Brain embodiment of syntax and grammar: Discrete combinatorial

Brain embodiment of syntax and grammar:
Discrete combinatorial mechanisms spelt out in neuronal circuits
Friedemann Pulvermüller
Medical Research Council, Cognition and Brain Sciences Unit, Cambridge CB2 7EF, UK
This paper has been accepted for publication and is in press in the journal
Brain and Language
running head: Brain Embodiment of Syntax
Key words: associative memory, brain, discrete combinatorial rule, grammar, neuronal
assembly, pushdown store, sequence detector, syntax, tree structure
234 words in abstract, 7,724 words in text, 155 references, 3 tables, 4 figures
Address for correspondence:
Friedemann Pulvermüller, Ph.D.
MRC Cognition and Brain Sciences Unit
15 Chaucer Road
Cambridge CB2 2EF, UK
phone: +44 1223 355294 ext. 670
fax: +44 1223 359062
e-mail: [email protected]
Acknowledgements: I would like to thank Valentino Braitenberg, Andreas Knoblauch,
Marion Ormandy, Yury Shtyrov, Mark Steedman, Kambiz Tavabi, and two anonymous
referees for their comments and help at different stages of this work. Supported by the
Medical Research Council (UK) (U1055.04.003.00001.01, U1055.04.003.00003.01) and by
the European Community under the “New and Emerging Science and Technologies”
Programme (NEST-2005-PATH-HUM contract 043374, NESTCOM).
2
BRAIN EMBODIMENT OF SYNTAX
ABSTRACT
Neuroscience has greatly improved our understanding of the brain mechanisms of abstract
lexical and semantic processes. The brain mechanisms underlying words and concepts are
distributed neuronal assemblies reaching into sensory and motor systems of the cortex and, at
the cognitive level, this mechanism is mirrored by the sensorimotor grounding of form and
meaning of symbols. Recent years have seen the emergence of evidence for similar brain
embodiment of syntax. Neurophysiological studies have accumulated support for the
linguistic notion of abstract combinatorial rules manifest as functionally discrete neuronal
assemblies. Concepts immanent to the theory of abstract automata could be grounded in
observations from modern neuroscience, so that it became possible to model abstract
pushdown storage – which is critical for building linguistic tree structure representations – as
ordered dynamics of memory circuits in the brain. At the same time, neurocomputational
research showed how sequence detectors already known from animal brains can be neuronally
linked so that they merge into larger functionally discrete units, thereby underpinning abstract
rule representations that syntactically bind lexicosemantic classes of morphemes and words
into larger meaningful constituents. Specific predictions of brain-based grammar models
could be confirmed by neurophysiological and brain imaging experiments using MEG, EEG
and fMRI. Neuroscience and neurocomputational research offering perspectives on
understanding abstract linguistic mechanisms in terms of neuronal circuits and their
interactions therefore point programmatic new ways to future theory-guided experimental
investigation of the brain basis of grammar.
3
BRAIN EMBODIMENT OF SYNTAX
1. Introduction
In spite of great progress in the neurosciences in understanding the mechanisms of
language and conceptual thought, there is still one domain that appears to be largely immune
to brain evidence. This domain, syntax, is far removed from concrete events and is built on
highly sophisticated abstraction. Even its most basic phenomena are best described in terms of
abstract formula and it therefore may appear doubtful – if not impossible – that its core
principles can be translated into the profane language of nerve cells and circuits. After all,
linguists have spent some effort to prove that it is impossible to build a finite neuronal
automaton processing relevant aspects of syntax (Chomsky, 1963) and the project of unifying
language and the brain at a theoretical level has therefore been long abandoned by most
language researchers. Paradoxically so, because the human brain is the very instrument whose
invention is equivalent with that of language. Important clues about the mechanisms of syntax
and grammar may therefore come from brain research in the effort to understand specifically
human abstract abilities. However, at present, the mainstream linguistic position seems to be
that neuroscience is still far from addressing the relevant questions (see, for example,
Jackendoff, 2002). To cite Chomsky: “The ERP (event-related potential) observations are just
curiosities, lacking a theoretical matrix” (Chomsky, 2000) (p. 25). Is this position wellfounded? This review will give the answer.
Linguists would not state that modern neuroscience ignores the topic of syntax.
Neuroimaging and also neuropsychology have in fact been immensely successful in defining
the brain regions activated during syntactic processing and even the regions whose lesion
leads to syntactic failure (Caplan, 2006; Friederici, 2002; Grodzinsky & Friederici, 2006;
Hagoort, 2005; Ullman, 2001). However, this work had so far little impact on linguistics. One
reason for this may be that much neuroscience research focused on the question of WHERE in
4
BRAIN EMBODIMENT OF SYNTAX
the brain syntax is localized, demonstrating that the relevant regions are mainly in left
perisylvian language cortex, but also involve adjacent areas, the insula and even widespread
areas in both the left and right hemispheres; and also subcortical structures are relevant,
including basal ganglia and thalamus. The where question as such is not considered to be
extremely exciting by many linguists, and skepticism towards modern variants of phrenology
(the 19th Century science of where aspects of the mind are located) is shared by
neuroscientists with a theoretical interest (for example, Knight, 2007).1 In contrast, the
question of WHEN physiological indicators of syntax would first occur does spark more
interest, especially in psycholinguists. The precise timing of EEG and MEG responses
provides clues about the time course of syntactic information access in the context of lexical
and semantic processing; this has implications for psycholinguistic theory (Friederici, 2002;
Hagoort, 2005, 2008; Pulvermüller & Shtyrov, 2006). However, the most interesting
functional question of HOW syntax is represented and computed in brain circuits, has, for a
long time, not been addressed at all.
At the mechanistic level of nerve cells and their connections, theoretical work on
syntax associated with empirical efforts is still relatively sparse. Substantial work focuses on
neural network architectures that may solve specific syntactic problems (Botvinick & Plaut,
2006; Chang, Dell, & Bock, 2006; Chater & Manning, 2006; Elman, 1990; Hanson &
Negishi, 2002) and some studies of neural networks even relate syntactic and serial order
computations to cortical areas and subcortical structures (Beiser & Houk, 1998; Dominey,
1
Note, furthermore, that the sluggish haemodynamic response revealed by fMRI is naturally difficult to relate to
rapid linguistic processes such as sentence comprehension. In many cases, experiments use secondary tasks and,
therefore, features of these tasks, rather than the comprehension of critical linguistic structures, may be reflected in
brain activation. Here is a citation from a recent publication which exemplifies and illustrates this problem: “Many
participants reported after the experiment that they had employed a strategy of internally generating a second
sentence of the form used in the comprehension task in order to prepare for the possible response. Thus, this kind of
task preparation may have engendered the frontoparietal activation observed here” (Bornkessel, Zysset, Friederici,
von Cramon, & Schlesewsky, 2005) (p. 230).
5
BRAIN EMBODIMENT OF SYNTAX
1997; Dominey, Hoen, Blanc, & Lelekov-Boissard, 2003; Voegtlin & Dominey, 2005).
However, only a few proposals exist that relate syntax to defined mechanisms, for example to
cell populations sequentially linked into synchronously firing “synfire” chains establishing
serial order, functionally discrete neuronal assemblies that structure the sequential activation
of other neuronal populations, and the ordered dynamics of whole brain areas brought about
by the various excitatory and inhibitory neuronal populations they house (Bienenstock, 1996;
Braitenberg, 1996; Hayon, Abeles, & Lehmann, 2005; Pulvermüller, 1993, 2003a;
Smolensky, 1999; Wennekers, Garagnani, & Pulvermüller, 2006; Wennekers & Palm, 1996;
Wickens, Hyland, & Anson, 1994). This paper will focus on the question what we know and
can reasonably assume about the mechanistic level of syntax in the brain. It will review views
on the embodiment of grammar in neuronal circuitry and discuss implications of neuroscience
experiments for linguistic theory.
2. Neuroscience evidence for phonological and semantic circuits
In neuroscience, recent progress led to a better understanding of the brain mechanisms
of phonology and semantics. The brain areas for phoneme processing were identified and
even fine-grained local differences in processing speech sounds with different features could
be mapped (Diesch, Eulitz, Hampson, & Ross, 1996; Obleser et al., 2006; Uppenkamp,
Johnsrude, Norris, Marslen-Wilson, & Patterson, 2006). Old models – according to which
language comprehension engages the posterior language area in the superior temporal cortex
of the dominant left hemisphere (“Wernicke’s area”) and the left inferior frontal language area
(“Broca’s area”) acts as a centre for speech production (Lichtheim, 1885; Wernicke, 1874) –
could be overthrown, especially by evidence for the involvement of inferiorfrontal and even
motor areas in language comprehension and, vice versa, the activation of temporal regions in
6
BRAIN EMBODIMENT OF SYNTAX
production (Fadiga, Craighero, Buccino, & Rizzolatti, 2002; Fadiga, Fogassi, Pavesi, &
Rizzolatti, 1995; Paus, Perry, Zatorre, Worsley, & Evans, 1996; Pulvermüller et al., 2006;
Pulvermüller, Shtyrov, & Ilmoniemi, 2003; Wilson, Saygin, Sereno, & Iacoboni, 2004;
Zatorre, Evans, Meyer, & Gjedde, 1992). Motor systems activation even reflects, and
contributes to, the perception of phonological features in the understanding of speech
(D’Ausilio et al., 2009; Pulvermüller et al., 2006). The picture emerging now is this:
Phonemes and morphemes, the smallest units that distinguish or carry meaning, are
represented and processed in the brain by distributed neuronal ensembles spread-out over
fronto-temporal language areas (Pulvermüller, 1999). These circuits are lateralized to the left
dominant hemisphere, although significant portions of their neurons may be housed in the
right non-dominant hemisphere (Grodzinsky & Friederici, 2006; Hickok & Poeppel, 2000;
Marslen-Wilson & Tyler, 2007; Pulvermüller & Mohr, 1996).
The links between signs and their meaning critically depend on long-distance corticocortical connections. Words referring to objects or actions correspond to lateralized frontotemporal circuits linking up with object or action representations in visual and other sensory
areas or motor systems of the brain (Barsalou, 2008; Martin, 2007; Pulvermüller, 2005). Clear
evidence for this view comes from neuroimaging studies and from work with patients. Frontal
areas activate specifically to words semantically related to actions, for example action verbs
such as ‘write’ or ‘speak’. Words semantically related to visually defined objects activate
specific parts of the inferior-temporal stream of object processing (Chao, Haxby, & Martin,
1999; Pulvermüller & Hauk, 2006; Simmons et al., 2007) and even words meaning odours
may activate specifically areas involved in olfaction (González et al., 2006). Hence, the brain
appears to classify words according to their referential semantic features and to the modalities
involved in processing this information. Semantic mappings – or semantic topographies – can
7
BRAIN EMBODIMENT OF SYNTAX
be so sharp that they even distinguish action words according to the body part with which the
referent action is preferably executed (Aziz-Zadeh, Wilson, Rizzolatti, & Iacoboni, 2006;
Buccino et al., 2005; Hauk, Johnsrude, & Pulvermüller, 2004; Pulvermüller, Härle, &
Hummel, 2000; Tettamanti et al., 2005). Critically, action-perception links may even
contribute to abstract semantic processing (Boulenger, Hauk, & Pulvermüller, 2009; Glenberg
& Kaschak, 2002; Glenberg, Sato, & Cattaneo, 2008).
The mentioned studies revealing activation loci – and therefore addressing the
WHERE question – have implications for cognitive theory. As motor-articulatory and
acoustic systems of the brain are co-activated in processing of language sounds, the results
endorse theories postulating a link of articulatory and acoustic-phonetic information at both
the neural and cognitive levels (Braitenberg & Pulvermüller, 1992; Fry, 1966; Galantucci,
Fowler, & Turvey, 2006; Liberman & Mattingly, 1985; Pulvermüller et al., 2006). At the
semantic level, the co-activation of left fronto-temporal language regions with modalityspecific areas reflecting semantic aspects of language materials supports neurocognitive
theories of embodied cognition that view the conceptual and semantic store as a result of
information exchange with the world in the context of perceptions and actions (Barsalou,
1999, 2008; Glenberg & Kaschak, 2002; Pulvermüller, 1999; Warrington & McCarthy, 1987).
Interestingly, these activations reflecting aspects of the meaning of words and sentences
emerge extremely rapidly, within 1/5 second after input information allows for unique
recognition of the critical word. This implies that, at the cognitive level, these activations
reflect language comprehension and meaning access – rather than late processes of
reprocessing or reanalysis (Hauk, Shtyrov, & Pulvermüller, 2008; Pulvermüller, 2005;
Pulvermüller, Shtyrov, & Ilmoniemi, 2005). Claims that the multimodal links in cortex might
not be essential for semantic and linguistic processing (Mahon & Caramazza, 2008) are
8
BRAIN EMBODIMENT OF SYNTAX
confronted with observations from the clinic: Word processing degrades in a category specific
manner if action- or visually-related brain areas are lesioned and a pronounced processing
deficit for action words and concepts accompanies degeneration of the motor system in Motor
Neuron Disease (Bak, O'Donovan, Xuereb, Boniface, & Hodges, 2001; Damasio, Grabowski,
Tranel, Hichwa, & Damasio, 1996; Neininger & Pulvermüller, 2003; Tranel, Damasio, &
Damasio, 1997). In Semantic Dementia, where frontotemporal areas are most severely
affected, the same subcategories of action- and visually-related words, which normally most
strongly activate these areas, are most severely impaired (Pulvermüller et al., 2009). These
lexico-semantic deficits accompanying circumscribed brain dysfunction endorse a critical role
of sensorimotor areas in the processing of form and meaning of symbols (Pulvermüller et al.,
2009).
3. Abstract combinatorial rules in the human brain: neurophysiological results
Notwithstanding the progress made in the phonological and the semantic domains, one
may judge that the core of our linguistic powers, the specifically human ability to variably
combine discrete meaningful units into rule-conforming higher-order structures, has so far not
been enlightened by neuroscience research (cf. Chomsky, 2000; Jackendoff, 2002). Still, in
the neurophysiological domain, effort has been spent to find brain indexes that possibly
indicate syntactic or grammatical processes. A good number of studies compare brain
responses to ungrammatical strings with those to grammatical, well-formed ones. This
comparison is particularly important to linguistic theory, as most theories of syntax provide
parses for grammatical strings, but separate out ungrammatical strings by parser blockage, a
distinction sometimes thought to be best assessable empirically using the grammaticality
9
BRAIN EMBODIMENT OF SYNTAX
judgment task. Neurophysiology may provide the tool for looking at the brain’s
grammaticality judgements directly.
A range of components of the event-related brain potential and field, including the
early left-anterior negativity (ELAN, Friederici, Pfeifer, & Hahne, 1993; ELAN, Neville,
Nicol, Barss, Forster, & Garrett, 1991), a late posterior positivity (P600, Hagoort, Brown, &
Groothusen, 1993; P600, Osterhout & Holcomb, 1992), and the early left-lateralised syntactic
Mismatch Negativity (sMMN, Shtyrov, Pulvermüller, Näätänen, & Ilmoniemi, 2003), are
enhanced to ungrammatical as compared with grammatical word strings (for review, see
Friederici, 2002; Hagoort, 2005; Kuperberg, 2007; Pulvermüller, Shtyrov, & Hauk, 2009).
Cortical localization indicates that two areas, the inferior frontal gyrus and the superiortemporal cortex, house the neuron populations that contribute to these physiological responses
potentially linked to grammatical sequence computation (Friederici, Wang, Herrmann, Maess,
& Oertel, 2000; Pulvermüller & Shtyrov, 2003; Shtyrov, Pulvermüller, Näätänen, &
Ilmoniemi, 2003). Interestingly, these same areas are also involved in processing nonlinguistic sequences (Fazio et al., 2009; Maess, Koelsch, Gunter, & Friederici, 2001).
However, it is difficult to decide whether the brain response to an ungrammatical
string reflects its syntactic features or rather the conditional probability with which words
follow each other. As every ungrammatical string is usually also a very rarely occurring item,
it is also possible that the degree of expectedness – or string probability – is reflected. This
point, which equally applies to comparisons between simple (and therefore frequent) and
more complex (rarer) sentence types (see discussion below), is especially virulent, as the
major brain responses discussed as possible reflections of ungrammaticality and syntactic
processing load – ELAN, P600 and sMMN – are similar in latencies and distributions to wellknown components of the event-related potential that reflect the probability with which
10
BRAIN EMBODIMENT OF SYNTAX
stimuli occur within an experiment. The ELAN, falls into the time range of the N100 and
N200 components, which are modulated by attention, the P600 resembles closely the
attention-related P300, and the sMMN is a left-lateralized variant of the Mismatch Negativity,
or MMN, which reflects change detection. Respectable arguments have been brought up to
defend the position that components elicited by ungrammatical sentences are different from
their nonlinguistic sisters – for example, that the syntactic P600 is different from the P300
(Osterhout & Hagoort, 1999) – but this does not prove a specifically syntactic role of the
respective component. It may still be that sequential probabilities in language use, rather than
the presence or absence of an underlying discrete combinatorial rule, determine the emergence
of a syntactic ELAN, P600 or MMN. Direct evidence for a neuronal correlate of rules is
therefore not provided by the work reviewed so far. Although the number of occurrences of
grammatical and ungrammatical strings can be balanced exactly within a given experiment,
the general occurrence probability of a sentence, sentence type, or relevant section thereof, is
usually higher than that of a similar asyntactic string. As both grammaticality and sequential
probability may have a correlate in the brain response, it is necessary to ask whether these
effects can be separated.
Important information about grammatical processes in the brain may come from
studies of different types of correct sentences differing in their grammatical structure, word
order or memory requirements. However, as performance shapes grammar (Hawkins, 1994), it
is naturally difficult to separate effects of grammatical structure per se from those of string
frequency. As already mentioned, if sentences with “canonical” and “noncanoncal” word
order are contrasted, the influence of string type frequencies may play a role. For example,
noncanoncical passive or object-relative sentences activated left inferiorfrontal/premotor and
superiortemporal areas more strongly than canonical active or subject-relative sentences
11
BRAIN EMBODIMENT OF SYNTAX
(Caplan et al., 2002; Just, Carpenter, Keller, Eddy, & Thulborn, 1996; Kinno, Kawamura,
Shioda, & Sakai, 2008). Similarly, the posterior inferiorfrontal area was more strongly active
for subordinate clauses with an initial accusative noun phrase (NP) referring to nonanimate
objects than for similar structures with an initial dative noun phrase including an animate
noun (Grewe et al., 2006); note that the former seem to be rarer than the latter and,
interestingly, grammaticality ratings reflect this frequency difference (Keller, 2001; Kempen
& Harbusch, 2003). Together with data on grammaticality violations mentioned above, these
results on grammatical sentence types are compatible with the idea that syntactic circuits are
housed in inferiorfrontal cortex, with an additional potential contribution of superiortemporal
areas. Interestingly, as Keller’s data demonstrate, a link between gradually more or less
canonical syntactic structure and grammaticality ratings exists (Keller, 2001). However, this
fact still leaves us with the critical question of how to separate string probability and
grammaticality per se.2
In a recent experiment using the sMMN as the dependent variable, an attempt was
made to dissociate rule- and probability-related processes. All critical sequences of words
were presented with equal probability in the experiment, but both grammaticality and the
probability with which these strings are used in language did vary. Critical words were
identical in different conditions so that lexical or semantic factors could not account for any
differences. Determiner-noun agreement was used as the syntactic feature of interest, as it
represents a syntactic link without semantic implication. As in previous work, grammatical
strings frequently used in language were compared with rare ungrammatical ones. However,
in a third condition grammatical but extremely rare strings were presented. The fourth logical
2
Further important issues are critical in the investigation of words, phrases and sentences with different
morphological and grammatical structure. It is important to match stimulus sentences for psycholinguistic variables
such as the length and standardized lexical frequency of their critical words, their sequential probability at their
12
BRAIN EMBODIMENT OF SYNTAX
possibility – ungrammatical strings that occur frequently – is not available, because a
descriptive grammar classifies frequently occurring strings as regular. In this sense, Bever and
colleagues pointed out that ungrammatical strings used in everyday language may initiate rule
changes (Bever, Carroll, & Hurtig, 1976). Because of their changing or even unclear
grammatical status, “frequent ungrammatical” strings were not included in the experiment.
For the third and critical condition, rare but grammatical strings, two competing
predictions are available. Cognitive scientists and computational modelers have claimed that
rules (in the sense of discrete combinatorial units) do not exist in neural systems, brains
included, and that probability mapping is the mechanism that explains the human ability to
string words in sentences (Elman et al., 1996; McClelland & Patterson, 2002). A probability
mapping approach without discrete neuronal units predicts that a rare grammatical string
elicits a brain response comparable to that of similarly rare ungrammatical strings. However,
if the brain uses a discrete rule mechanism to distinguish between grammatical and
ungrammatical strings, this mechanism should be involved to the same degree in the
processing of rare and common grammatical word strings with the same syntactic structure,
but it should fail to process ungrammatical word sequences. This does not rule out a possible
physiological difference between brain processes sparked by rare and common grammatical
strings.3 Therefore, probability mapping and rule access theories lead to different predictions
on the brain response to rare grammatical strings of words.
The syntactic enhancement of the MMN was found to rare ungrammatical strings
compared with common grammatical ones, as had been reported in a range of previous
experiments (Hasting, Kotz, & Friederici, 2007; Menning et al., 2005; Pulvermüller, Shtyrov,
position in the stimulus sentence or sentence type, their cloze probabilities and other features of the context along with
lexical properties (Pulvermüller, 2007; Pulvermüller & Shtyrov, 2006).
13
BRAIN EMBODIMENT OF SYNTAX
Hasting, & Carlyon, 2008; Shtyrov, Pulvermüller, Näätänen, & Ilmoniemi, 2003). Relevant
cortical generators are being sparked early (60-200 ms) in inferiorfrontal and superiortemporal
cortex. Critically, syntactic but rare phrases led to an MMN similar to that of common
grammatical strings (Figure 1). The sMMN seems to behave as if it was an automatic
grammaticality index (Pulvermüller, Shtyrov, Hasting, & Carlyon, 2008). The
neurophysiological rule correlate revelaed by the sMMN provides support for the rule access
theory (Pulvermüller & Assadollahi, 2007) and is consistent with the existence of discrete
combinatorial neuronal processors generating the discrete neurophysiological responses.
Rules appear to be effective in the brain’s syntactic computations.
Figure 1 about here
4. Recursion, pushdown storage and syntactic linkage
A simple sentence, such as (1) in Table 1, can be assembled from words, morphemes
and phrases by successively applying rewriting rules, for example (2a)-(e). Together with an
additional rule, (3), (2a) can be applied repeatedly, so that the string becomes longer and the
same syntactic relationship is established between newly added items. This describes the wellknown recursivity (or recursiveness) of grammar. Generally, a function is recursive if the
function itself is applied within its own definition. Rules (2a) and (3) together therefore result
in recursivity.
Insert Table 1 about here
Linguists have argued that languages spoken by humans all have the feature of
recursivity, allowing us, in principle, to build infinitely many sentences out of a limited
3
As the brain is so sensitive to conditional probabilities (Donchin, 1981; Kutas & Hillyard, 1984; Näätänen, Gaillard, &
Mäntysalo, 1978; Näätänen, Tervaniemi, Sussman, Paavilainen, & Winkler, 2001), it is not surprising to find a physiological
correlate for them in an experiment investigating word sequences.
14
BRAIN EMBODIMENT OF SYNTAX
vocabulary and set of rules. Languages include a few 10,000s of morphemes (Pinker, 1994)
and current grammar theories use in the range of 100 rule formula to describe relevant
fragments of their grammars (see, for example, Gazdar, Klein, Pullum, & Sag, 1985).
Linguists sometimes argue that, in principle, indefinitely many sentences can thus be built
from a limited set of elements, although this statement assumes that strings can be endless,
and under this assumption even the duck language is indefinitely recursive, as sequences of
“quack”s could also be considered to be, in principle, not limited in length. Clearly, indefinite
recursiveness cannot therefore be a critical feature of human language.
However, recursion of a structurally interesting type, especially the variant making it
possible to embed sentences into each other – as illustrated by examples (1), (4) and (5) above
– might be characteristic of humans (Hauser, Chomsky, & Fitch, 2002). The duck’s endless
quacking is easily generated by recursive activity in a finite state automaton (Kleene, 1956;
McCulloch & Pitts, 1943), whereas the complex sentences are not (Chomsky, 1963). It is
therefore of the essence to ask what algorithms and, ultimately, brain mechanisms such
special recursion is based on.
Chomsky showed that the generation of center-embedded strings of the type discussed
requires a context-free grammar (including rule pairs like (2a) and (3)) and that this type of
grammar can be mechanistically implemented as one kind of abstract automaton (Chomsky,
1963). The automaton type does not need the great computational capacity of a Turing
machine – which uses a store, in which it can, at any point in time, freely access all its
memory entries. Chomsky showed that a machine with reduced processing power and
memory access, which reads from and writes into its store in a predefined order, is
advantageous for processing language, especially certain types of complex sentences
(Hopcroft, Motwani, & Ullman, 2001). The last entry into the store has to be retrieved first,
15
BRAIN EMBODIMENT OF SYNTAX
the second second, and the first entry entered is retrieved first. As this ordering is similar to
the one generated by putting items on a stack and later removing them again (from top to
bottom), the terms “pushdown” or “stack automaton” are sometimes used. In the above
example, with each application of the rules (2a) and (3), a new VP symbols is put on the stack,
“pushing down” any symbols that might already sit on the stack. In retrieval, the last VP
entered pops up first, calling for immediate completion of the phrase started last and so on.
This iterative procedure results in a centre-embedded structure (cf. sentences (6)-(8)).
Availability of multiple stacks and the possibility to move items between stacks further
increases the generation power of the automaton (Joshi & Schabes, 1997).
5. Is there a neurophysiological basis for pushdown storage?
Memory researchers found nerve cells that store specific memory contents (Fuster,
1997, 2003). If a monkey has to keep in memory that a stimulus (characterized, for example,
by a specific colour) has been presented and the animal has to find a matching stimulus later,
cells which are specific to memory content (the relevant colour) become strongly active.
Many of these cells show strong activation initially and subsequently loose activity
continuously – following approximately an exponential or power function. Interestingly, these
memory cells would stay at an enhanced state of activity until the sample stimulus appears
and the memory content is being retrieved. It has been argued that the mechanism underlying
the continuous activity of memory cells is the reverberation of neuronal excitation in
distributed neuronal assemblies or memory circuits (Fuster, 1997, 2003; Fuster & Jervey,
1981). Circuits of this type may take a role in object or action representation and also in
controlling sequences of processing events (e.g., Sigala, Kusunoki, Nimmo-Smith, Gaffan, &
Duncan, 2008).
16
BRAIN EMBODIMENT OF SYNTAX
The neurophysiological mechanisms revealed by memory cell activity can provide the
brain basis of pushdown storage. Given there are several memory circuits whose dynamics are
governed by the same activation-deactivation function, these circuits can be activated in
sequence and will each start loosing activity thereafter. As the first circuit will therefore have
lost activity when the second is sparked, and both of these will have (further) declined when a
third one activates, the hierarchy of circuit activation levels reflects the temporal order of
inputs. The circuit activated last will be at the highest level and the first-activated will be at
the lowest, resembling a pushdown stack. A read-out mechanism only accessing the most
highly active circuit reads the last memory entry first and the first entry last, thus providing
typical pushdown memory retrieval (Figure 2, Braitenberg & Pulvermüller, 1992;
Pulvermüller, 1993). Multiple pushdown stacks can, in principle, be located in different brain
areas. Neuroscience research therefore suggests a brain correlate of abstract devices critical
for representing and processing complex sentences.
Figure 2 about here
Interestingly, single linguistic representations may be used several times in the
processing of a sentence. This is a requirement for recursion. Note that, in the generation of
example sentence (5), the N and V symbols are each used three times and three “copies” of
the VP symbol – for the complements of the three initial NPs – remain in the pushdown store
for some time. Linguists have claimed that neural networks cannot cope with this problem of
multiple processing of the same element, which was dubbed the “problem of two”, “type
token problem”, or “problem of multiple instantiation” (Jackendoff, 2002; Marcus, 2001;
Sougné, 1998). This critique seems, however, to be aimed at neural networks of a certain kind
(e.g., Elman et al., 1996; e.g., Rumelhart & McClelland, 1987). Applied to the brain itself, it
17
BRAIN EMBODIMENT OF SYNTAX
would lead to the apparent paradox that the language organ of humans cannot process human
language – an irrational position.
Whereas a large class of neural networks is unable to cope with the problem of
multiple instantiation, solutions are offered by specific brain-inspired network architectures
(Shastri & Ajjanagadde, 1993; Sougné, 1998, 2001; van der Velde & de Kamps, 2006). One
mechanism for the multiple processing of individual representations is provided by neuronal
circuits that can hold several distinct activations at a time. Such distinct activations can be
precisely timed synchronous waves traveling independently in the same discrete circuit, for
example in reverberatory synfire chains (for details and simulations, see Hayon, Abeles, &
Lehmann, 2005). There is evidence for synfire chains from neuroscience experiments (Abeles,
Bergman, Margalit, & Vaadia, 1993; Plenz & Thiagarajan, 2007), although the proposal of a
co-existence of two or three traveling waves of activity in the same representation is presently
based on computer simulations. Together, multiple circuit activation and neuronal pushdown
storage provide critical neural mechanisms that may underpin the binding of linguistic
constituents in syntactic strings (Pulvermüller, 2003a). However, both proposals are at a
theoretical, neurocomputational level, still awaiting direct experimental support.
As the outlined brain basis of abstract algorithmic syntactic processes is consistent
with current knowledge about brain function, it would be false to state that linguistic
representations and processes are far removed from brain mechanisms. It will be critical for
future investigations of syntax to have available a brain model of syntax that predicts brain
activation based on the dynamics of syntactic circuits. Availability of such models is
necessary for bridging the brain-mind gap in the language sciences.
Abstract algorithmic rules can be applied iteratively, without an upper limit for the
number of iterations. This limitlessness had been an argument for the superiority of grammar
18
BRAIN EMBODIMENT OF SYNTAX
systems to neural networks, which always have restricted numbers of elements and were
therefore believed not to be capable of generating context-free grammars equivalent to
pushdown automata, let alone mildly context-sensitive ones and devices employing multiple
stacks. Linguists had previously argued that neural architectures can be modified to
accommodate a pushdown mechanism (e.g., Petri, 1970; Schnelle, 1996). The present
proposal is also open to limitless iterations. Both the number of simultaneously active circuits
(holding activity at different levels) and the number of simultaneously reverberating activation
waves in a given circuit can, in principle, be increased without limit. This allows free choice
of the numbers of embedding levels and instantiations of a given category. However, a
realistic neuronal model of syntax should account for limits also, as it aims to realistically
mirror brain mechanisms, not abstract ones. As a matter of fact, strings with two or three
levels of centre-embedding are documented in language use, but are extremely rare (Karlsson,
2007) and, if used at all, frequently misunderstood (Bach, Brown, & Marslen-Wilson, 1986).
Noise leading to merging of simultaneously traveling waves in the same circuit accounts for
difficulties caused by word repetitions, for example the fact that sentence (9) is more difficult
to process than (4) – although these sentences have the same syntactic structure (see also
Marcus, 2008). It seems important to offer accounts of these and similar limitations within a
neuronal language theory (Christiansen & Chater, 1999).
Taken together, these considerations related to system limitations and noise suggest
that the neuronal pushdown store and multiple activations of the same neuronal assembly are
fragile mechanisms that should be used sparsely in grammar theories. While a neuronal model
of grammar may exploit them in the processing of sentence embedding, a less fragile and
complementary system would be advantageous for computing syntactic relationships within
19
BRAIN EMBODIMENT OF SYNTAX
simple sentences and clauses. Is there evidence for more elementary mechanisms of syntactic
linkage?
6. From sequence detectors in animals to syntactic linkage
Morphemes, words and larger constituents within a sentence are related to each other.
These relationships are not restricted to elements that directly follow each other in the
sequence. Linear models of syntactic links, describing pairs, triplets or, more generally, ktuplets of adjacent morphemes, are therefore insufficient for syntactic analysis. Sentence (10)
(Figure 3) illustrates some local and distant syntactic links in a sentence.
Figure 3 about here
There are next-neighbor syntactic links between the transitive verb and its complements, the
head noun of the subject and the object realized as pronoun. Also, the verb and its affix can be
analyzed as morphologically linked. Non-local syntactic relationships include agreement
between subject noun and verb suffix (Kate … -s) and verb-particle linkage (build … up).
Describing these relationships in a two-dimensional tree structure is not easily possible,
because projection lines would cross if the links between connected but distant constituents
were added to the tree (dashed lines in (10b), Figure 3). Linguistic theories did therefore
introduce additional mechanisms – for example percolation of grammatical features through
the tree (for an introduction, see Haegeman, 1991) – to capture noun-verb agreement or
movement to account for particle displacement. However, after a slight revision of the
mechanism for syntactic linkage, these additional mechanisms may not be necessary.
Syntactic links between meaningful language units, words or larger constituents,
including the ones described by rewriting rules ((2), (3)), agreement and other non-local
syntactic relationships, have been proposed to be neurobiologically grounded in discrete
20
BRAIN EMBODIMENT OF SYNTAX
combinatorial neuronal assemblies, DCNAs, that link together pairs of constituents
(Pulvermüller, 2003a). The DCNAs are discrete as their activation takes one of several
discrete states, they are combinatorial as they link together, or bind, specific lexical and
morphological categories, and they are higher-order neuronal representations of linguistic
units above the level of lexical units. Apart from linking categories together, typical DCNAs
establish a temporal order between the category members they bind to.4 In the neuronal
framework, each of the syntactic links indicated by lines in (10a) is carried by one DCNA.
Is there reason for assuming DCNAs? The central nervous system of higher mammals
(Barlow & Levick, 1965; Hubel, 1995) and already that of invertebrates (Reichardt & Varju,
1959; Varju & Reichardt, 1967) includes neuronal units, so-called motion detectors, that
respond specifically to sequences. The sequence detectors link to two other neuronal units, A
and B, and respond specifically if these “input” units are activated in a given order, A–then–
B, but not, or much less, to the reverse order of activations, B–then–A. The mechanisms
resulting in sequence-sensitivity are diverse and may involve non-linear computations (Borst,
2007; Egelhaaf, Borst, & Reichardt, 1989), but may also result from an additive integration
process (Hubel, 1995). As sequence sensitive units are so common in nervous systems, it
would be surprising if they were unavailable for processing combinatorial information in
sentences. Sequence detectors for morphemes and words may provide a mechanism for
dynamically linking constituents pair-wise, thus realizing syntactic links that are effective in
processing a sentence (Pulvermüller, 2003b). The linkage includes the previous activation of
first and second word in a string (e.g., noun-verb), the resultant ignition of the order-sensitive
DCNA and a binding process functionally linking the latter with the active lexical
representations. The binding mechanism is best thought of as synchronized oscillatory activity
4
DCNAs that do not impose temporal order (thus acting, in principle, as AND units for two constituents) are
21
BRAIN EMBODIMENT OF SYNTAX
at high frequencies (gamma band), as oscillatory dynamics appear to reflect lexical (Canolty
et al., 2007; Pulvermüller, Birbaumer, Lutzenberger, & Mohr, 1997) and sentence processing
(Bastiaansen & Hagoort, 2006; Weiss et al., 2005). A critical question remains how abstract
syntactic processing units operating on classes of lexical items can be built from sequence
detectors (Section 7).
Note that the non-local nature of the syntactic binding mechanism5, which may apply
to adjacent elements or units that are separated in time, does not make it necessary to postulate
a principal difference between phrase structure relationships between adjacent words in a
sentence, as captured by tree structures, and agreement mechanisms between distant units, nor
does it run into difficulty in the processing and representation of discontinuous constituents,
such as particle verbs (see example sentence (10)). These few remarks on the perspectives of
neurobiological grammar mechanisms must remain sketchy, but a neuronal grammar model
integrating neuronal pushdown storage (Section 5) and syntactic linkage by way of DCNAs
has been laid out elsewhere (Pulvermüller, 2003a). In this approach, the elementary clauseand sentence-internal level of processing is dealt with by syntactic sequence detectors and the
more complex simultaneous and recursive processing of more than one sentence, for example
in the case of embedding, is dealt with by the pushdown memory component.
7. Discrete combinatorial neuronal assemblies: How they emerge
Are syntax and grammar largely inborn or rather learned? As mentioned in Section 3,
neuronal network modelers view probability mapping in simple recurrent networks as
thought to underlie free constituent order and certain forms of scrambling (Pulvermüller, 2003a).
5
The term “binding” has a range of different usages in linguistics. Binding sometimes specifically refers to the referential
properties of nominals (Chomsky, 1981), or to a range of different syntactic relationships that lead to unification of
constituents (Hagoort, 2005). Here, the term “syntactic binding” is used in the latter, widest sense, further assuming that a
neurophysiological binding mechanisms by oscillatory synchrony provides the neuronal mechanism for it (Pulvermüller,
Birbaumer, Lutzenberger, & Mohr, 1997).
22
BRAIN EMBODIMENT OF SYNTAX
sufficient for growing syntax in automata (see Section 3, Elman et al., 1996; see Section 3,
McClelland & Patterson, 2002; Rumelhart & McClelland, 1987). The linguistic notion of a
discrete combinatorial rule is, in this view, ill-defined and syntactic regularities can be learned
by associative learning principles. Some linguists, on the other hand, deny the relevance of
learning for building syntactic representations. In their view, the genetic code provides the
critical machinery and learning plays a peripheral role – comparable to the labeling of the
keys of a typewriter whose mechanistic system is already fully in place (Chomsky, 1980).
However, research into statistical language learning shows that a lot of structural information
about syntax and grammar can be extracted from corpora and therefore does not require preprogrammed knowledge. For example, information about the lexical classification of words
can be extracted from the local context in which they appear using statistical (Briscoe &
Carroll, 1997; Lin, 1998; Townsend & Bever, 2001) or neural network methods (Elman,
1990; Honkela, Pulkki, & Kohonen, 1995). If genuinely grammatical information can be
extracted by the learner from corpora and language input, the postulate of inborn information
about lexical categories violates the parsimony principle and the preferable solution is to
attribute the formation of lexico-syntactic categories to learning.
On the other hand, an unconstrained associative learning perspective not providing any
neural basis for discrete combinatorial rules appears to be insufficient too, as the rulelessness
assumption seems to be in conflict with current neuroscience evidence (Section 3). In essence,
the linguistic approach appears correct in postulating rules, whereas neural networks
perspectives rightly exploit statistical information about correlation of lexical items. The
critical question now becomes whether a putative brain correlate of discrete combinatorial
rules might emerge in brain-like networks due to biologically plausible correlation learning.
23
BRAIN EMBODIMENT OF SYNTAX
The concept of DCNAs for syntactic binding (cf. footnote 5) offers a new perspective
on this debate. Neuronal modeling encourages precise questions about which specific types of
wirings are a prerequisite of grammar and about the precise role of associative learning. A
possible mechanism is this: Elementary sequence detectors (Section 6) are sensitive to ordered
pairs of words/morphemes, and would, by way of reciprocal connections and synchronization,
link functionally (or merge) with two lexical circuits, thereby building a higher-order syntactic
unit. Obviously, “binding units” for specific word pairs are insufficient for processing the
structure of a sentence, because classes of lexical elements, not individual items, need to be
linked together. A fruitful idea here is that a range of binding units for word pairs become
linked among each other. The emerging complex aggregates of sequence detectors thereby
become sensitive to any lexical element included in a particular syntactic class or lexical
category followed by any other element belonging to a different lexical category (Knoblauch
& Pulvermüller, 2005; Pulvermüller, 2003b).
In a simulation project exploring syntactic learning, we used a pre-structured autoassociative memory with built-in sequence detectors for every possible pair sequence of
words. The “grammar area” of the network did not only include these pre-wired sequence
detectors, but, in addition, initially very weak links between all pairs of sequence detectors.
This was done in an attempt to incorporate an important feature of the cortex (O'Reilly, 1998).
Based on the statistics of cortical connections, it has been argued that the cortex is, in fact, an
auto-associative memory (Braitenberg & Schüz, 1998) and a network with similar auto-rich
auto-associative connectivity (Palm, 1993; Willshaw, Buneman, & Longuet-Higgins, 1969)
was therefore chosen as a basis of the modeling. Binding of sequence detectors into circuits
operating on classes of lexical items was indeed observed in these simulations. This binding
was critically related to the availability of sequence detectors and auto-associative links
24
BRAIN EMBODIMENT OF SYNTAX
between them, which could be strengthened by associative learning. As Figure 4 shows,
sequence detectors in the network’s grammar area (black dots) became strongly linked not
only to their respective lexical representations (red lines to grey dots at the left and top) but
also among each other (black lines). The links between the elementary sequence detectors
sensitive to word pairs were selectively strengthened because sequence detectors were
regularly co-activated in the process of string learning and word substitution between strings.
This co-activation of sequence detectors led to the emergence of neuronal aggregates
comprising sequence detectors sensitive to similar contexts. Because of the strong internal
connections within the neuronal aggregates, they acted as a higher-order discrete functional
unit for syntactic binding.
Figure 4 and Table 2 about here
Note that the network explains generalization by a simple mechanism. Any possible
first word is bound into the syntactic binding circuit, as any possible second item is, too.
Activity will therefore spread from any active neuronal unit included in the first lexical class
to all items included in class 2, regardless of whether the specific sequence had been learned
before or not. Even though only some of the possible noun-verb pairs had been learned,
sequence learning and substitution between a subset of the nouns and verbs in their respective
contexts led to priming between any noun and verb included into corresponding combinatorial
subsets (Table 2). Subsequent to priming by possible first string elements, full activation of
any of the possible second elements led to DCNA ignition and to syntactic binding of the
constituent pair involved.
25
BRAIN EMBODIMENT OF SYNTAX
In one sense, one can argue that, after learning, syntactic binding units formed in the
network provided a mechanistic correlate of an abstract binary rule.6 Interestingly,
combinatorial information taken from the British National Corpus, BNC, led to subclassification of nouns and verbs into lexico-semantic sub-categories and sub-rules. Figure 4
shows separable sub-rules/DCNAs for nouns denoting human subjects and verbs referring to
specifically human actions, and nouns referring to flying objects and verbs related to flying.
These syntactic-semantic sub-classifications depended on the threshold with which the
network operates: At a low activation threshold, the two DCNA merged into one yielding
syntactic priming between all nouns and verbs. At a higher threshold, the syntactic-semantic
rules were kept separate. The functionally discrete rule representations could also overlap to a
degree neuroanatomically, that is, in the sequence detectors they consisted of, while still
staying functionally separate (for further discussion, see Pulvermüller & Knoblauch, 2009).
Grades of canonicity and grammaticality of constructions can be represented by DCNAs with
different connection strengths. Note that the neuronal binding mechanism illustrated for single
words can also operate on larger syntactic constituents.
In sum, DCNAs that unify lexico-semantic categories and develop as outlined above
can act as syntactic binding circuits in establishing grammatical relationships in simple
sentences (see Figure 3). Although not all syntactic phenomena are easily modeled by these
emerging aggregates of sequence detectors, they do provide a candidate neuronal mechanism
for one type of grammar rule (Knoblauch & Pulvermüller, 2005; Pulvermüller, 2003b;
Pulvermüller & Knoblauch, 2009).
6
Here is a description of the type of rule covered: Let a and b be two syntactic categories, where members of a
usually precedes b members. The rule SD(ab) licences the n*m sequences of elements of a and b, and does so even though
26
BRAIN EMBODIMENT OF SYNTAX
8.
Syntax embedded in semantics: Relationship to Cognitive and Construction Grammar
Whereas most grammar frameworks list syntactic rules, the lexicon and semantics as
separate components, the brain-mechanistic framework views DCNAs as directly connected
to word (morpheme, and other lexical) representations, which, in turn, connect with their
semantic networks. Syntax and semantics are therefore functionally connected and the
“syntactic” DCNAs, after ratifying a given order of constituents, can mediate the activation
equilibrium between lexical-semantic circuits. Also, as illustrated in Section 7, DCNAs bind
together lexical-semantic classes of constituents (see Table 2, Figure 4). Lexical-semantic
word categories are manifest in the sets of DCNAs they connect with. These units, in turn,
determine the grammatical strings in which the lexical units regularly occur. This
amalgamated nature of syntax and semantics puts the neurobiological perspective in the
vicinity of lexical grammar theories in the cognitive linguistic tradition viewing grammar as
an integration machinery for form and meaning (e.g., Bresnan, 2001; Lakoff, 1987;
Langacker, 1991).
The neurobiological-mechanistic approach assumes (a) basic morpheme
representations linking form and meaning and, in addition, (b) storage of representations for
morpheme sequences by way of sequence detectors and discrete combinatorial neuronal
assemblies, DCNAs (see also Feldman, 2006; Feldman & Narayanan, 2004; see also
Pulvermüller, 2003a). This view fits well with proposals in the Construction Grammar
framework that word sequences can be represented, or “listed”, as higher-order lexical items
(Goldberg, 2006; Lakoff, 1987). The present proposal takes this idea further and views rules
as emergent properties of multiple listings of sequences, bound together as a consequence of
the recombination of string segments. In addition, a distinction is proposed between the
the input is sparse, that is, the number of learned strings is substantially below n*m. Such a rule becomes recursive if the
27
BRAIN EMBODIMENT OF SYNTAX
separate listing, or representation, of 1/ morpheme sequences and 2/ single higher-order
lexical items, complex words or idioms. Different neurophysiological indexes may reflect
linguistic binding at the lexical and syntactic levels (Shtyrov, Pihko, & Pulvermüller, 2005;
Shtyrov, Pulvermüller, Näätänen, & Ilmoniemi, 2003).
As mentioned, syntactic DCNAs unify word-related circuits that bind information
about word form and meaning (Section 2), and therefore, there is direct interaction between
syntactic binding and the dynamics of semantic circuits. This may provide a neurobiological
perspectives on semantic compositionality, as the stronger or weaker links from the syntactic
binding circuit to specific lexical units have a direct impact on semantic activation. This can
be illustrated using compound words. For the compound noun swine flue, the DCNA
connection would activate flue more strongly than swine, so that the full set of semantic
features of the lexicosemantic representation of the main part of the compound (or head) is
kept activate upon initial excitation. The (modifier) noun swine, whose circuit is connected to
the DCNA through weaker links, would receive less priming, thus modeling the contextspecific attribute character of this item. When recognizing flue swine, however, the opposite
dynamics account for the understanding of “swine with flue as an attribute”. In languages like
French, where modifiers tend to follow heads, the links of the DCNAs established by learning
would impose the reverse activation hierarchy. In case the meaning of compounds and larger
structures is not easily derived from the semantics of their parts, semantic circuits can be
linked to sequence detectors and lexical representations of strings and constructions (Section
2). This may be relevant for implementing the meaning of complex forms that are
semantically opaque and cannot be derived compositionally (cash cow). It is needless to say
that the various semantic relationships that can hold between the different parts of a
pushdown mechanism of section 5 is functional.
28
BRAIN EMBODIMENT OF SYNTAX
compound expression may require the involvement of additional semantic circuits and
schemas (Feldman, 2006; Lakoff, 1987). The mechanisms outlined here may help to
neurobiologically underpin the tight link between syntax and semantics postulated in the
context of Cognitive and Construction Grammar.
9. Strong auto-associative links as critical step in language evolution?
The present approach assumes a division of labor between DCNAs and neuronal
pushdown mechanisms (Table 3). Subtypes of syntactic mechanisms, supporting more local
vs. non-local processing, may relate to different brain areas (Opitz & Friederici, 2007),
although not all studies support such a spatial dissociation (Indefrey et al., 2001). Independent
of any difference in cortical localization, the two proposed mechanisms supported by DCNAs
and pushdown store are fundamentally different at the functional level. However, in spite of
their functional differences, they may rely on similar brain-structural requirements. Critical
for the development of DCNAs is a high degree of auto-associative connectivity in a
“grammar area”, between sequence specific units. A high degree of auto-associative
connectivity may also be critical for the constant ignition and decline of activation in different
memory circuits, which is essential for the neuronal pushdown store (Braitenberg &
Pulvermüller, 1992). If connections within circuits are extremely strong, activity decrease
may become the consequence of habituation phenomena shared by large quantities of
excitatory neurons and thus also shared by different circuits. It may therefore be that rich
auto-associative connectivity between the main language areas in left-perisylvian cortex is a
requirement for the human capacity to process syntax and grammar. Apart from local
neuroanatomical specificity in this area (e.g., Harasty, Seldon, Chan, Halliday, & Harding,
2003; Hutsler & Galuske, 2003; e.g., Jacobs et al., 1993), the surprisingly strongly developed
29
BRAIN EMBODIMENT OF SYNTAX
connections between left-superiortemporal and -inferiorfrontal areas in humans (but not other
primates) may be of relevance here (Catani, Jones, & Ffytche, 2005; Rilling et al., 2008; Saur
et al., 2008). These strong links may enable the human cortex to build especially strongly
connected syntactic circuits that form the basis of a pushdown store, as they may allow the
binding of string-specific sequence detectors into generalizing DCNA, thereby providing a
basic rule mechanism. Models and data therefore seem to converge on the conclusion that a
high degree of auto-associative connectivity in left-perisylvian language cortex might have
been the critical evolutionary step towards human language.
Insert Table 3 about here
10. Summary and Conclusion
This paper reviewed some abstract properties of neuronal networks in general and the
human brain in particular. First, the so-called rule debate was mentioned and new evidence
from neurophysiological experiments for the existence of abstract syntactic rule
representations in the human brain was discussed. The claim that rules exist as brain
mechanisms calls for concrete mechanistic and brain-inspired models of syntactic and
grammatical processing. A brain model of recursion and embedding was reviewed and related
to abstract structural approaches to grammar. The interim conclusions were that abstract
automata and brains provide equivalent pushdown mechanisms for recursive processes and
that progress can be achieved in tailoring the limitations of these devices to those of real
brains. Sequence detectors and sequence detector aggregates, DCNAs, were discussed as a
brain basis for syntactic linkage of morphemes and larger constituents and the interplay
between circuits possibly built-in into specific structures of the human brain and associative
learning was put in the context of neuronal rule formation. The two proposed brain-internal
30
BRAIN EMBODIMENT OF SYNTAX
grammar mechanisms were related to each other (Table 3) and it was hypothesized that the
emergence of strong auto-associative connectivity in the perisylvian language cortex was the
critical step in human language evolution. Throughout, relationships between grammar
theories and neuronal approaches were mentioned.
31
BRAIN EMBODIMENT OF SYNTAX
References
Abeles, M., Bergman, H., Margalit, E., & Vaadia, E. (1993). Spatiotemporal firing patterns in the frontal cortex
of behaving monkeys. Journal of Neurophysiology, 70, 1629-1638.
Aziz-Zadeh, L., Wilson, S. M., Rizzolatti, G., & Iacoboni, M. (2006). Congruent embodied representations for
visually presented actions and linguistic phrases describing actions. Curr Biol, 16(18), 1818-1823.
Bach, E., Brown, C., & Marslen-Wilson, W. (1986). Crossed and nested dependencies in German and Dutch: a
psycholinguistic study. Language and Cognitive Processes, 1, 249-262.
Bak, T. H., O'Donovan, D. G., Xuereb, J. H., Boniface, S., & Hodges, J. R. (2001). Selective impairment of verb
processing associated with pathological changes in Brodmann areas 44 and 45 in the Motor Neurone
Disease-Dementia-Aphasia syndrome. Brain, 124, 103-120.
Barlow, H., & Levick, W. R. (1965). The mechanism of directionally selective units in rabbit's retina. Journal of
Physiology, 178, 477-504.
Barsalou, L. W. (1999). Perceptual symbol systems. Behav Brain Sci, 22(4), 577-609; discussion 610-560.
Barsalou, L. W. (2008). Grounded cognition. Annu Rev Psychol, 59, 617-645.
Bastiaansen, M., & Hagoort, P. (2006). Oscillatory neuronal dynamics during language comprehension. Prog
Brain Res, 159, 179-196.
Beiser, D. G., & Houk, J. C. (1998). Model of cortical-basal ganglionic processing: encoding the serial order of
sensory events. J Neurophysiol, 79(6), 3168-3188.
Bever, T. G., Carroll, J. M., & Hurtig, M. (1976). Analogy or Ungrammatical sequences that are utterable and
comprehensible are the origin of new grammars in language acquisition and linguistic evolution. In T.
G. Bever, J. J. Katz & D. T. Langendoen (Eds.), An integrated theory of linguistic ability. New York:
T.Y. Crowell Press.
Bienenstock, E. (1996). On the dimensionality of cortical graphs. Journal of Physiology (Paris), 90, 251-256.
Bornkessel, I., Zysset, S., Friederici, A. D., von Cramon, D. Y., & Schlesewsky, M. (2005). Who did what to
whom? The neural basis of argument hierarchies during language comprehension. Neuroimage, 26(1),
221-233.
Borst, A. (2007). Correlation versus gradient type motion detectors: the pros and cons. Philos Trans R Soc Lond
B Biol Sci, 362(1479), 369-374.
Botvinick, M. M., & Plaut, D. C. (2006). Short-term memory for serial order: a recurrent neural network model.
Psychol Rev, 113(2), 201-233.
Boulenger, V., Hauk, O., & Pulvermüller, F. (2009). Grasping ideas with the motor system: Semantic
somatotopy in idiom comprehension. Cereb Cortex, in press.
Braitenberg, V. (1996). Il gusto della lingua. Meccanismi cerebrali e strutture grammaticali. Merano: Alpha &
Beta.
Braitenberg, V., & Pulvermüller, F. (1992). Entwurf einer neurologischen Theorie der Sprache.
Naturwissenschaften, 79, 103-117.
Braitenberg, V., & Schüz, A. (1998). Cortex: statistics and geometry of neuronal connectivity (2 ed.). Berlin:
Springer.
Bresnan, J. (2001). Lexical Functional Syntax. Oxford: Blackwell Publishers Ltd.
Briscoe, E., & Carroll, J. (1997). Automatic extraction of subcategorization from corpora. Proceedings of the 5th
ACL Conference on Applied Natural Language Processing, Washington, DC, 5, 1-9.
Buccino, G., Riggio, L., Melli, G., Binkofski, F., Gallese, V., & Rizzolatti, G. (2005). Listening to action-related
sentences modulates the activity of the motor system: a combined TMS and behavioral study. Brain Res
Cogn Brain Res, 24(3), 355-363.
Canolty, R. T., Soltani, M., Dalal, S. S., Edwards, E., Dronkers, N. F., Nagarajan, S. S., Kirsch, H. E., Barbaro,
N. M., & Knight, R. T. (2007). Spatiotemporal dynamics of word processing in the human brain. Front.
Neurosci., doi:10.3389/neuro.01.1.1.014.2007.
Caplan, D. (2006). Aphasic deficits in syntactic processing. Cortex, 42(6), 797-804.
Caplan, D., Vijayan, S., Kuperberg, G., West, C., Waters, G., Greve, D., & Dale, A. M. (2002). Vascular
responses to syntactic processing: event-related fMRI study of relative clauses. Hum Brain Mapp, 15(1),
26-38.
Catani, M., Jones, D. K., & Ffytche, D. H. (2005). Perisylvian language networks of the human brain. Annals of
Neurology, 57(1), 8-16.
Chang, F., Dell, G. S., & Bock, K. (2006). Becoming syntactic. Psychol Rev, 113(2), 234-272.
32
BRAIN EMBODIMENT OF SYNTAX
Chao, L. L., Haxby, J. V., & Martin, A. (1999). Attribute-based neural substrates in temporal cortex for
perceiving and knowing about objects. Nature Neuroscience, 2(10), 913-919.
Chater, N., & Manning, C. D. (2006). Probabilistic models of language processing and acquisition. Trends in
Cognitive Sciences, 10(7), 335-344.
Chomsky, N. (1963). Formal properties of grammars. In R. D. Luce, R. R. Bush & E. Galanter (Eds.), Handbook
of mathematical psychology. Volume 2 (pp. 323-418). New York, London: Wiley.
Chomsky, N. (1980). Rules and representations. New York: Columbia University Press.
Chomsky, N. (1981). Lectures on government and binding. Dordrecht: Foris.
Chomsky, N. (2000). New horizons in the study of language and mind. Cambridge: Cambridge University Press.
Christiansen, M. H., & Chater, N. (1999). Toward a connectionist model of recursion in human linguistic
performance. Cognitive Science, 23(2), 157-205.
D’Ausilio, A., Pulvermüller, F., Salmas, P., Bufalari, I., Begliomini, C., & Fadiga, L. (2009). The Motor
Somatotopy of Speech Perception. Current Biology, in press.
Damasio, H., Grabowski, T. J., Tranel, D., Hichwa, R. D., & Damasio, A. R. (1996). A neural basis for lexical
retrieval. Nature, 380, 499-505.
Diesch, E., Eulitz, C., Hampson, S., & Ross, B. (1996). The neurotopography of vowels as mirrored by evoked
magnetic field measurements. Brain Lang, 53(2), 143-168.
Dominey, P. F. (1997). An anatomically structured sensory-motor sequence learning system displays some
general linguistic capacities. Brain Lang, 59(1), 50-75.
Dominey, P. F., Hoen, M., Blanc, J. M., & Lelekov-Boissard, T. (2003). Neurological basis of language and
sequential cognition: evidence from simulation, aphasia, and ERP studies. Brain Lang, 86(2), 207-225.
Donchin, E. (1981). Surprise! ... Surprise? Psychophysiology, 18, 493-513.
Egelhaaf, M., Borst, A., & Reichardt, W. (1989). Computational structure of a biological motion-detection
system as revealed by local detector analysis in the fly's nervous system. Journal of the Optical Society
of America (A), 6, 1070-1087.
Elman, J. L. (1990). Finding structure in time. Cognitive Science, 14, 179-211.
Elman, J. L., Bates, L., Johnson, M., Karmiloff-Smith, A., Parisi, D., & Plunkett, K. (1996). Rethinking
innateness. A connectionist perspective on development. Cambridge, MA: MIT Press.
Fadiga, L., Craighero, L., Buccino, G., & Rizzolatti, G. (2002). Speech listening specifically modulates the
excitability of tongue muscles: a TMS study. European Journal of Neuroscience, 15(2), 399-402.
Fadiga, L., Fogassi, L., Pavesi, G., & Rizzolatti, G. (1995). Motor facilitation during action observation: a
magnetic stimulation study. J Neurophysiol, 73(6), 2608-2611.
Fazio, P., A., C., Craighero, L., D’Ausilio, A., Roy, A., Pozzo, T., Calzolari, F., Granieri, E., & Fadiga, L.
(2009). Encoding of human action in Broca's area. Brain, in press.
Feldman, J. (2006). From molecule to metaphor: A neural theory of language. Cambridge, MA: MIT Press.
Feldman, J., & Narayanan, S. (2004). Embodied meaning in a neural theory of language. Brain and Language,
89(2), 385-392.
Friederici, A. D. (2002). Towards a neural basis of auditory sentence processing. Trends in Cognitive Sciences,
6(2), 78-84.
Friederici, A. D., Pfeifer, E., & Hahne, A. (1993). Event-related brain potentials during natural speech
processing: effects of semantic, morphological and syntactic violations. Cognitive Brain Research, 1(3),
183-192.
Friederici, A. D., Wang, Y., Herrmann, C. S., Maess, B., & Oertel, U. (2000). Localization of early syntactic
processes in frontal and temporal cortical areas: a magnetoencephalographic study. Hum Brain Mapp,
11(1), 1-11.
Fry, D. B. (1966). The development of the phonological system in the normal and deaf child. In F. Smith & G.
A. Miller (Eds.), The genesis of language (pp. 187-206). Cambridge, MA: MIT Press.
Fuster, J. M. (1997). Network memory. Trends in Neurosciences, 20(10), 451-459.
Fuster, J. M. (2003). Cortex and mind: Unifying cognition. Oxford: Oxford University Press.
Fuster, J. M., & Jervey, J. P. (1981). Inferotemporal neurons distinguish and retain behaviorally relevant features
of visual stimuli. Science, 212, 952-955.
Galantucci, B., Fowler, C. A., & Turvey, M. T. (2006). The motor theory of speech perception reviewed.
Psychon Bull Rev, 13(3), 361-377.
Gazdar, G., Klein, E., Pullum, G., & Sag, I. (1985). Generalized phrase structure grammar. Cambridge, MA:
Harvard University Press.
Glenberg, A. M., & Kaschak, M. P. (2002). Grounding language in action. Psychonomic Bulletin & Review, 9(3),
558-565.
33
BRAIN EMBODIMENT OF SYNTAX
Glenberg, A. M., Sato, M., & Cattaneo, L. (2008). Use-induced motor plasticity affects the processing of abstract
and concrete language. Curr Biol, 18(7), R290-291.
Goldberg, A. E. (2006). Constructions at work: The nature of generalisation in language. Oxford: Oxford
University Press.
González, J., Barros-Loscertales, A., Pulvermüller, F., Meseguer, V., Sanjuán, A., Belloch, V., & Ávila, C.
(2006). Reading cinnamon activates olfactory brain regions. Neuroimage, 32(2), 906-912.
Grewe, T., Bornkessel, I., Zysset, S., Wiese, R., von Cramon, D. Y., & Schlesewsky, M. (2006). Linguistic
prominence and Broca's area: the influence of animacy as a linearization principle. Neuroimage, 32(3),
1395-1402.
Grodzinsky, Y., & Friederici, A. D. (2006). Neuroimaging of syntax and syntactic processing. Curr Opin
Neurobiol, 16(2), 240-246.
Haegeman, L. (1991). Introduction to government and binding theory. Cambridge, MA: Basil Blackwell.
Hagoort, P. (2005). On Broca, brain, and binding: a new framework. Trends Cogn Sci, 9(9), 416-423.
Hagoort, P. (2008). The fractionation of spoken language understanding by measuring electrical and magnetic
brain signals. Philos Trans R Soc Lond B Biol Sci, 363(1493), 1055-1069.
Hagoort, P., Brown, C., & Groothusen, J. (1993). The syntactic positive shift (SPS) as an ERP-measure of
syntactic processing. Language and Cognitive Processes, 8(4), 439-483.
Hanson, S. J., & Negishi, M. (2002). On the emergence of rules in neural networks. Neural Comput, 14(9), 22452268.
Harasty, J., Seldon, H. L., Chan, P., Halliday, G., & Harding, A. (2003). The left human speech-processing
cortex is thinner but longer than the right. Laterality, 8(3), 247-260.
Hasting, A. S., Kotz, S. A., & Friederici, A. D. (2007). Setting the stage for automatic syntax processing: the
mismatch negativity as an indicator of syntactic priming. J Cogn Neurosci, 19(3), 386-400.
Hauk, O., Johnsrude, I., & Pulvermüller, F. (2004). Somatotopic representation of action words in the motor and
premotor cortex. Neuron, 41, 301-307.
Hauk, O., Shtyrov, Y., & Pulvermüller, F. (2008). The time course of action and action-word comprehension in
the human brain as revealed by neurophysiology. J Physiol Paris, 102(1-3), 50-58.
Hauser, M. D., Chomsky, N., & Fitch, W. T. (2002). The faculty of language: what is it, who has it, and how did
it evolve? Science, 298(5598), 1569-1579.
Hawkins, J. (1994). A performance theory of order and constituency (Vol. 73). Cambridge: Cambridge
University Press.
Hayon, G., Abeles, M., & Lehmann, D. (2005). A model for representing the dynamics of a system of synfire
chains. J Comput Neurosci, 18(1), 41-53.
Hickok, G., & Poeppel, D. (2000). Towards a functional neuroanatomy of speech perception. Trends in
Cognitive Sciences, 4(4), 131-138.
Honkela, T., Pulkki, V., & Kohonen, T. (1995). Contextual relations of words in grimm tales analyzed by selforganizing map. proceedings of international conference on artificial neural networks. In F. FogelmanSoulie' & P. Gallinari (Eds.), Proceedings of the International Conference on Neural Networks
(ICANN) 1995 (pp. 3-7). Paris: EC2 et Cie.
Hopcroft, J. E., Motwani, R., & Ullman, J. D. (2001). Introduction to automata theory, languages, and
computation. New York: ACM.
Hubel, D. (1995). Eye, brain, and vision (2 ed.). New York: Scientific American Library.
Hutsler, J., & Galuske, R. A. (2003). Hemispheric asymmetries in cerebral cortical networks. Trends Neurosci,
26(8), 429-435.
Indefrey, P., Brown, C. M., Hellwig, F., Amunts, K., Herzog, H., Seitz, R. J., & Hagoort, P. (2001). A neural
correlate of syntactic encoding during speech production. Proceedings of the National Academy of
Sciences,USA, 98(10), 5933-5936.
Jackendoff, R. (2002). Foundations of language: Brain, meaning, grammar, evolution. Oxford, UK: Oxford
University Press.
Jacobs, B., Batal, H. A., Lynch, B., Ojemann, G., Ojemann, L. M., & Scheibel, A. B. (1993). Quantitative
dendritic and spine analyses of speech cortices: a case study. Brain and Language, 44, 239-253.
Joshi, A. K., & Schabes, Y. (1997). Tree-adjoining gammars. In A. Salomon & E. Rozenberg (Eds.), Handbook
of formal languages and automata. Berlin: Springer-Verlag.
Just, M. A., Carpenter, P. A., Keller, T. A., Eddy, W. F., & Thulborn, K. R. (1996). Brain activation modulated
by sentence comprehension. Science, 274(5284), 114-116.
Karlsson, F. (2007). Constraints on multiple centre embedding of clauses. Journal of Linguistics (2007), 43 :
365-392 Cambridge University Press 43, 365-392.
34
BRAIN EMBODIMENT OF SYNTAX
Keller, F. (2001). Gradience in grammar: experimental and computational aspects of degrees of grammaticality.
University of Edinburgh, Edinburgh.
Kempen, G., & Harbusch, K. (2003). An artificial opposition between grammaticality and frequency: comment
on Bornkessel, Schlesewsky, and Friederici (2002). Cognition, 90(2), 205-210; discussion 211-203,
215.
Kinno, R., Kawamura, M., Shioda, S., & Sakai, K. L. (2008). Neural correlates of noncanonical syntactic
processing revealed by a picture-sentence matching task. Hum Brain Mapp, 29(9), 1015-1027.
Kleene, S. C. (1956). Representation of events in nerve nets and finite automata. In C. E. Shannon & J.
McCarthy (Eds.), Automata studies (pp. 3-41). Princeton, NJ: Princeton University Press.
Knight, R. T. (2007). Neuroscience. Neural networks debunk phrenology. Science, 316(5831), 1578-1579.
Knoblauch, A., & Pulvermüller, F. (2005). Sequence detector networks and associative learning of grammatical
categories. In S. Wermter, G. Palm & M. Elshaw (Eds.), Biomimetic neural learning for intelligent
robots (pp. 31-53). Berlin: Springer.
Kuperberg, G. R. (2007). Neural mechanisms of language comprehension: challenges to syntax. Brain Res, 1146,
23-49.
Kutas, M., & Hillyard, S. A. (1984). Brain potentials during reading reflect word expectancy and semantic
association. Nature, 307, 161-163.
Lakoff, G. (1987). Women, fire, and dangerous things. What categories reveal about the mind. Chicago:
University of Chicago Press.
Langacker, R. (1991). Foundations of Cognitive Grammar. 2 vols. Stanford: Stanford University Press.
Liberman, A. M., & Mattingly, I. G. (1985). The motor theory of speech perception revised. Cognition, 21, 1-36.
Lichtheim, L. (1885). On aphasia. Brain, 7, 433-484.
Lin, D. (1998). Automatic retrieval and clustering of similar words. Proceedings of the17th International
Conference on Computational Linguistics and 36th Annual Meeting of the Association for
Computational Linguistics, Montreal, 17, 768-773.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in Broca's area: an
MEG study. Nat Neurosci, 4(5), 540-545.
Mahon, B. Z., & Caramazza, A. (2008). A critical look at the embodied cognition hypothesis and a new proposal
for grounding conceptual content. J Physiol Paris, 102(1-3), 59-70.
Marcus, G. F. (2001). The algebraic mind: integrating connectionism and cognitive science. Cambridge, MA:
MIT Press.
Marcus, G. F. (2008). Kluge: The haphazard construction of the human mind. Boston, MA: Houghton Mifflin
Co.
Marslen-Wilson, W. D., & Tyler, L. K. (2007). Morphology, language and the brain: the decompositional
substrate for language comprehension. Philos Trans R Soc Lond B Biol Sci, 362(1481), 823-836.
Martin, A. (2007). The representation of object concepts in the brain. Annu Rev Psychol, 58, 25-45.
McClelland, J. L., & Patterson, K. (2002). Rules or connections in past-tense inflections: what does the evidence
rule out? Trends Cogn Sci, 6(11), 465-472.
McCulloch, W. S., & Pitts, W. H. (1943). A logical calculus of ideas immanent in nervous activity. Bulletin of
Mathematical Biophysics, 5, 115-133.
Menning, H., Zwitserlood, P., Schoning, S., Hihn, H., Bolte, J., Dobel, C., Mathiak, K., & Lutkenhoner, B.
(2005). Pre-attentive detection of syntactic and semantic errors. Neuroreport, 16(1), 77-80.
Näätänen, R., Gaillard, A. W., & Mäntysalo, S. (1978). Early selective-attention effect on evoked potential
reinterpreted. Acta Psychologica, 42, 313-329.
Näätänen, R., Tervaniemi, M., Sussman, E., Paavilainen, P., & Winkler, I. (2001). 'Primitive intelligence' in the
auditory cortex. Trends in Neurosciences, 24(5), 283-288.
Neininger, B., & Pulvermüller, F. (2003). Word-category specific deficits after lesions in the right hemisphere.
Neuropsychologia, 41(1), 53-70.
Neville, H., Nicol, J. L., Barss, A., Forster, K. I., & Garrett, M. F. (1991). Syntactically based sentence
processing classes: evidence from event-related brain potentials. Journal of Cognitive Neuroscience, 3,
151-165.
O'Reilly, R. C. (1998). Six principles for biologically based computational models of cortical cognition. Trends
Cogn Sci, 2(11), 455-562.
Obleser, J., Boecker, H., Drzezga, A., Haslinger, B., Hennenlotter, A., Roettinger, M., Eulitz, C., & Rauschecker,
J. P. (2006). Vowel sound extraction in anterior superior temporal cortex. Hum Brain Mapp, 27(7), 562571.
35
BRAIN EMBODIMENT OF SYNTAX
Opitz, B., & Friederici, A. D. (2007). Neural basis of processing sequential and hierarchical syntactic structures.
Hum Brain Mapp, 28(7), 585-592.
Osterhout, L., & Hagoort, P. (1999). A superficial resemblance does not necessarily mean you are part of the
family: Counterarguments to Coulson, King and Kutas (1998) in the P600/SPS-P300 debate. Language
and Cognitive Processes, 14, 1-14.
Osterhout, L., & Holcomb, P. J. (1992). Event-related brain potentials elicited by syntactic anomaly. Journal of
Memory and Language, 31, 785-806.
Palm, G. (1993). On the internal structure of cell assemblies. In A. Aertsen (Ed.), Brain theory: spatio-temporal
aspects of brain function (pp. 261-270). Amsterdam: Elsevier.
Paus, T., Perry, D. W., Zatorre, R. J., Worsley, K. J., & Evans, A. C. (1996). Modulation of cerebral blood flow
in the human auditory cortex during speech: role of motor-to-sensory discharges. European Journal of
Neuroscience, 8(11), 2236-2246.
Petri, C. A. (1970). Kommunikation mit Automaten. Dissertation: Universität Bonn.
Pinker, S. (1994). The language instinct. How the mind creates language. New York: Harper Collins Publishers.
Plenz, D., & Thiagarajan, T. C. (2007). The organizing principles of neuronal avalanches: cell assemblies in the
cortex? Trends Neurosci, 30(3), 101-110.
Pulvermüller, F. (1993). On connecting syntax and the brain. In A. Aertsen (Ed.), Brain theory - spatio-temporal
aspects of brain function (pp. 131-145). New York: Elsevier.
Pulvermüller, F. (1999). Words in the brain's language. Behavioral and Brain Sciences, 22, 253-336.
Pulvermüller, F. (2003a). The neuroscience of language. Cambridge: Cambridge University Press.
Pulvermüller, F. (2003b). Sequence detectors as a basis of grammar in the brain. Theory in Biosciences, 122, 87103.
Pulvermüller, F. (2005). Brain mechanisms linking language and action. Nature Reviews Neuroscience, 6(7),
576-582.
Pulvermüller, F. (2007). Word processing in the brain as revealed by neurophysiological imaging using EEG and
MEG. In G. Gaskell (Ed.), Handbook of Psycholinguistics (pp. 119-140). Oxford: Oxford University
Press.
Pulvermüller, F., & Assadollahi, R. (2007). Grammar or serial order?: Discrete combinatorial brain mechanisms
reflected by the syntactic Mismatch Negativity. Journal of Cognitive Neuroscience, 19(6), 971-980.
Pulvermüller, F., Birbaumer, N., Lutzenberger, W., & Mohr, B. (1997). High-frequency brain activity: its
possible role in attention, perception and language processing. Progress in Neurobiology, 52(5), 427445.
Pulvermüller, F., Cooper-Pye, E., Dine, C., Hauk, O., Nestor, P., & Patterson, K. (2009). The word processing
deficit in Semantic Dementia: All categories are equal but some categories are more equal than others.
Journal of Cognitive Neuroscience in press.
Pulvermüller, F., Härle, M., & Hummel, F. (2000). Neurophysiological distinction of verb categories.
Neuroreport, 11(12), 2789-2793.
Pulvermüller, F., & Hauk, O. (2006). Category-specific processing of color and form words in left frontotemporal cortex. Cerebral Cortex, 16(8), 1193-1201.
Pulvermüller, F., Huss, M., Kherif, F., Moscoso del Prado Martin, F., Hauk, O., & Shtyrov, Y. (2006). Motor
cortex maps articulatory features of speech sounds. Proceedings of the National Academy of Sciences,
USA, 103(20), 7865-7870.
Pulvermüller, F., & Knoblauch, A. (2009). Discrete combinatorial circuits emerging in neural networks: A
mechanism for rules of grammar in the human brain? Neural Netw, 22(2), 161-172.
Pulvermüller, F., & Mohr, B. (1996). The concept of transcortical cell assemblies: a key to the understanding of
cortical lateralization and interhemispheric interaction. Neuroscience and Biobehavioral Reviews, 20,
557-566.
Pulvermüller, F., & Shtyrov, Y. (2003). Automatic processing of grammar in the human brain as revealed by the
mismatch negativity. Neuroimage, 20, 159-172.
Pulvermüller, F., & Shtyrov, Y. (2006). Language outside the focus of attention: the mismatch negativity as a
tool for studying higher cognitive processes. Progress in Neurobiology, 79(1), 49-71.
Pulvermüller, F., Shtyrov, Y., Hasting, A., & Carlyon, R. P. (2008). Syntax as a reflex: Neurophysiological
evidence for early automaticity of grammatical processing. Brain and Language, 104(1), 244-253.
Pulvermüller, F., Shtyrov, Y., & Hauk, O. (2009). Understanding in an instant: Neurophysiological evidence for
mechanistic language circuits in the brain. Brain and Language, in press.
Pulvermüller, F., Shtyrov, Y., & Ilmoniemi, R. J. (2003). Spatio-temporal patterns of neural language
processing: an MEG study using Minimum-Norm Current Estimates. Neuroimage, 20, 1020-1025.
36
BRAIN EMBODIMENT OF SYNTAX
Pulvermüller, F., Shtyrov, Y., & Ilmoniemi, R. J. (2005). Brain signatures of meaning access in action word
recognition. Journal of Cognitive Neuroscience, 17(6), 884-892.
Reichardt, W., & Varju, D. (1959). Übertragungseigenschaften im Auswertesystem für das Bewegungssehen.
Zeitschrift für Naturforschung, 14b, 674-689.
Rilling, J. K., Glasser, M. F., Preuss, T. M., Ma, X., Zhao, T., Hu, X., & Behrens, T. E. (2008). The evolution of
the arcuate fasciculus revealed with comparative DTI. Nat Neurosci, 11(4), 426-428.
Rumelhart, D. E., & McClelland, J. L. (1987). Learning the past tense of English verbs: implicit rules or parallel
distributed processing. In B. MacWhinney (Ed.), Mechanisms of language acquisition. Hillsdale, NJ:
Erlbaum.
Saur, D., Kreher, B. W., Schnell, S., Kummerer, D., Kellmeyer, P., Vry, M. S., Umarova, R., Musso, M.,
Glauche, V., Abel, S., Huber, W., Rijntjes, M., Hennig, J., & Weiller, C. (2008). Ventral and dorsal
pathways for language. Proc Natl Acad Sci U S A, 105(46), 18035-18040.
Schnelle, H. (1996). Die Natur der Sprache. Die Dynamik der Prozesse des Sprechens und Verstehens (2 ed.).
Berlin, New York: Walter de Gruyter.
Shastri, L., & Ajjanagadde, V. (1993). From simple associations to systematic reasoning: a connectionist
representation of rules, variables and dynamic bindings using temporal synchrony. Behavioral and
Brain Sciences, 16, 417-494.
Shtyrov, Y., Pihko, E., & Pulvermüller, F. (2005). Determinants of dominance: Is language laterality explained
by physical or linguistic features of speech? Neuroimage, 27(1), 37-47.
Shtyrov, Y., Pulvermüller, F., Näätänen, R., & Ilmoniemi, R. J. (2003). Grammar processing outside the focus of
attention: an MEG study. Journal of Cognitive Neuroscience, 15(8), 1195-1206.
Sigala, N., Kusunoki, M., Nimmo-Smith, I., Gaffan, D., & Duncan, J. (2008). Hierarchical coding for sequential
task events in the monkey prefrontal cortex. Proc Natl Acad Sci U S A, 105(33), 11969-11974.
Simmons, W. K., Ramjee, V., Beauchamp, M. S., McRae, K., Martin, A., & Barsalou, L. W. (2007). A common
neural substrate for perceiving and knowing about color. Neuropsychologia, 45(12), 2802-2810.
Smolensky, P. (1999). Grammar-based connectionist approaches to language. Cognitive Science, 23(4), 589-613.
Sougné, J. (1998). Connectionism and the problem of multiple instantiation. Trends in Cognitive Sciences, 2,
183-189.
Sougné, J. (2001). Binding and multiple instantiation in a distributed network of spiking nodes Connection
Science, 13(2), 99-126.
Tettamanti, M., Buccino, G., Saccuman, M. C., Gallese, V., Danna, M., Scifo, P., Fazio, F., Rizzolatti, G.,
Cappa, S. F., & Perani, D. (2005). Listening to action-related sentences activates fronto-parietal motor
circuits. Journal of Cognitive Neuroscience, 17(2), 273-281.
Townsend, D. J., & Bever, T. G. (2001). Sentence comprehension. Cambridge, MA: MIT Press.
Tranel, D., Damasio, H., & Damasio, A. R. (1997). A neural substrate for the retrieval of conceptual knowledge.
Neuropsychologia, 35, 1319-1327.
Ullman, M. T. (2001). A neurocognitive perspective on language: the declarative/procedural model. Nat Rev
Neurosci, 2(10), 717-726.
Uppenkamp, S., Johnsrude, I. S., Norris, D., Marslen-Wilson, W., & Patterson, R. D. (2006). Locating the initial
stages of speech-sound processing in human temporal cortex. Neuroimage, 31(3), 1284-1296.
van der Velde, F., & de Kamps, M. (2006). Neural blackboard architectures of combinatorial structures in
cognition. Behav Brain Sci, 29(1), 37-70; discussion 70-108.
Varju, D., & Reichardt, W. (1967). Übertragungseigenschaften im Auswertesystem für das Bewegungssehen II.
Zeitschrift für Naturforschung, 22b, 1343-1351.
Voegtlin, T., & Dominey, P. F. (2005). Linear recursive distributed representations. Neural Netw, 18(7), 878895.
Warrington, E. K., & McCarthy, R. A. (1987). Categories of knowledge: further fractionations and an attempted
integration. Brain, 110, 1273-1296.
Weiss, S., Mueller, H. M., Schack, B., King, J. W., Kutas, M., & Rappelsberger, P. (2005). Increased neuronal
communication accompanying sentence comprehension. Int J Psychophysiol, 57(2), 129-141.
Wennekers, T., Garagnani, M., & Pulvermüller, F. (2006). Language models based on Hebbian cell assemblies. J
Physiol Paris, 100, 16-30.
Wennekers, T., & Palm, G. (1996). Controlling the speed of synfire chains (Vol. 1112/1996). Berlin: Springer.
Wernicke, C. (1874). Der aphasische Symptomencomplex. Eine psychologische Studie auf anatomischer Basis.
Breslau: Kohn und Weigert.
Wickens, J., Hyland, B., & Anson, G. (1994). Cortical cell assemblies: a possible mechanism for motor
programs. J Mot Behav, 26(2), 66-82.
37
BRAIN EMBODIMENT OF SYNTAX
Willshaw, D. J., Buneman, O. P., & Longuet-Higgins, H. C. (1969). Non-holographic associative memory.
Nature, 222(197), 960-962.
Wilson, S. M., Saygin, A. P., Sereno, M. I., & Iacoboni, M. (2004). Listening to speech activates motor areas
involved in speech production. Nat Neurosci, 7(7), 701-702.
Zatorre, R. J., Evans, A. C., Meyer, E., & Gjedde, A. (1992). Lateralization of phonetic and pitch discremination
in speech processing. Science, 256, 846-849.
38
BRAIN EMBODIMENT OF SYNTAX
TABLE LEGENDS
Table 1: Example sentences and syntactic rewriting rules for illustrating recursion and
embedding. For explanation, see text. N–noun, NP–noun phrase, S–sentence, V–verb, VP–
verb phrase, Æ-is rewritten as, curly brackets index sets of lexical categories.
Table 2: The merging of sequence detectors into a neuronal aggregate (see Figure 4) leads to
priming between syntactic-semantic groups of lexical elements. The table shows the
physiological effect of such priming as revealed by a network simulation. The numbers give
the activation values of second string elements, verbs, (at the top) after activation of first
string elements, nouns (left). Note that, after activation of first string elements (nouns)
involved in substitutions in similar contexts (verbs), activation levels of context elements also
involved in the substitutions were enhanced to similar degrees (green areas). The neuronal
aggregates provide a generalized functional connection between subsets of nouns and verbs
also linking new pairs the network had never encountered together.
Table 3: Dual syntactic mechanism model: Different roles of syntactic discrete combinatorial
neuronal assemblies (DCNAs) and neuronal pushdown store in providing the mechanistic
brain basis for syntax and grammar.
39
BRAIN EMBODIMENT OF SYNTAX
FIGURE CAPTIONS
Figure 1: The syntactic Mismatch Negativity (MMN) occurred to rare ungrammatical strings
(black curve and bar), but not to similarly rare grammatical sentences (in red) or to frequently
occurring grammatical strings (in green and blue). Event-related magnetic MMN responses
are shown at the top and their statistics at the bottom. The syntactic MMN therefore reflected
grammaticality, but not string probability (Pulvermüller & Assadollahi, 2007).
Figure 2: Sequential activation of several memory circuits with graceful activity decline
(colored curves) yields storage of the activation order in the hierarchy of activity levels of the
assemblies. At a later time point (dashed line), the circuit sparked first will be at the lowest
activity level and the assembly activated last will still be at the highest level. This mechanism
can act as a neuronal implementation of a pushdown store (modified from Pulvermüller,
1993).
Figure 3: Syntactic links between constituents of a simple example sentence, (10). Diagram
(a) represents links implemented by syntactic binding circuits (DCNAs). Diagram (b) presents
a syntactic tree structure resulting from application of rewriting rules, plus syntactic
relationships not easily captured by trees (broken lines).
Figure 4: Combinatorial information immanent to noun-verb sequences and network result of
learning this information in an auto-associative memory including sequence detectors for
word pairs. The diagram on the left presents the matrix of co-occurrences and substitutions of
20 nouns and verbs obtained from the British National Corpus, BNC (in number of word pair
40
BRAIN EMBODIMENT OF SYNTAX
occurrences per 100 million words). The diagram on the right shows the network of lexical
circuits and sequence detectors and the connections strengthened by learning the string set
obtained from BNC. Grey circles in the periphery represent neural units corresponding to the
words in the diagram on the left. The central matrix shows sequence detectors corresponding
to word pair sequences. Filled black circles indicate sequence detectors whose respective
word pair was in the input, leading to strengthening of connections between sequence detector
and word representations (in red). Black lines show strengthened auto-associative links
between pairs of sequence detectors. At the top left and bottom right, two discrete
combinatorial neuronal assemblies, DCNAs, have formed. Depending on the threshold of
activation, these DCNAs either bind all nouns to verbs, or provide specific syntactic-semantic
linkage of action verbs and nouns related to living entities, and of flight-related verbs and
flying-object nouns (after Pulvermüller & Knoblauch, 2009).
41
BRAIN EMBODIMENT OF SYNTAX
Table 1
42
BRAIN EMBODIMENT OF SYNTAX
Table 2:
43
BRAIN EMBODIMENT OF SYNTAX
Table 3
44
BRAIN EMBODIMENT OF SYNTAX
Figure 1
45
BRAIN EMBODIMENT OF SYNTAX
Figure 2
46
BRAIN EMBODIMENT OF SYNTAX
Figure 3
47
BRAIN EMBODIMENT OF SYNTAX
Figure 4