c2--------must be first

Information Binding with
Dynamic Associative Representations
ARAKAWA, Naoya
Imaging Science and Engineering Laboratory, Tokyo Institute of Technology
Abstract. This paper proposes an explanation of human cognitive functions in terms of association. In this proposal, representation is not static
but is achieved by dynamic retrieval of patterns by means of association. With this dynamic representation, information binding in cognitive
functions such as physical world recognition, planning and language understanding is illustrated.
Keywords: association, generativity, dynamic representation, the binding problem
1
Introduction
This paper addresses the issue of representation in human cognitive functions
in terms of association and attempts to show the benefit of seeing cognitive
representation as dynamic associative processes with examples of perception,
planning and language understanding. In this paper, association designates the
process of obtaining a pattern from another through learning. As discussed in
[4], patterns here can be sub-symbolic vector representations formed by nonsupervised learning.
The reason why association is taken to be the basis for explaining cognitive functions here is two-fold: 1) cognitive functions are often accounted for
with sub-symbolic information processing and association is considered to be
its abstraction; 2) while sub-symbolic information processing is formalized with
models such as neural networks and Bayesian models, its technical details could
obscure the essence of over-all functions in certain discussions. Having said that
association is an abstraction of sub-symbolic information processing, the following are reasons why sub-symbolic information processing is important for
explaining cognitive functions. First, for models of brain functions such as neural networks are normally assumed to be sub-symbolic, their incorporation would
be essential for biologically natural or realistic cognitive modeling. Second, while
learning is an essential cognitive function, most modern-day learning models are
sub-symbolic. Finally, sub-symbolic information processing would evade thorny
issues in cognitive science such as the issue of classical categories brought forth by
cognitive linguists such as Lakoff [7] and the frame problem. Associative memory
█2
could evade the issue of classical categories, since sub-symbolic associative models (e.g., neural models) can represent prototypical and fuzzy categories. Certain
frame problems for intelligent agents to find relevant information could also be
evaded, as sub-symbolic associative memory could retrieve most relevant information (association) for given patterns first (by means of a competitive process
among‘ neurons, ’for example) (see [5] for a debate).
However, there are numerous problems with using association with cognitive
modeling. First of all, it is not clear if association alone is enough for explaining
dynamic cognitive functions such as planning or language understanding, as
association is nothing more than mapping between patterns. Secondly, as it was
argued in the computationalism vs. connectionism debate (see [2]), it is not
clear how connectionist/associationist models account for generativity of, say,
linguistic structure. Moreover, associationism also bears the so-called binding
problem for explaining the integration of various pieces of information (see [12]
and 2.1 below).
Given the aforementioned pros and cons, my motivation is to explain human
cognitive functions as an associative processes by addressing the problems raised
above.
2
The Proposal
The basic idea of the proposal is that associative representation is better conceived dynamically as traversal of association paths, if we are to address the
issues of information binding and generativity. Note that the issues are not
problematic in the case of classical symbolic models (e.g., [9]) or symbolic/subsymbolic hybrid cognitive architectures (e.g., [1] [10]), in which the composition/binding of information can be done symbolically by means of variable binding. However, when we regard the human brain as a sub-symbolic associative
machine, the apparent lack of binding mechanism becomes problematic for explaining human cognitive functions. Therefore, a (sub-) motivation here is to
explain cognitive functions with an associative mechanism without bringing in
variable binding. While the motivation is mainly explanatory, the proposal may
entertain technological merit: a cognitive system may be constructed with subsymbolic learning mechanisms without generating symbolic rules for the sake
of simplicity, although whether such a system has a practical advantage would
depend on the case.
In the sections to follow, the proposed method will be explained more in
detail by using concrete examples. First, physical world (scene) recognition is
discussed as the basic case of the binding problem. Then, discussions of planning
and language understanding will follow.
2.1
Physical world recognition
Recognizing the physical environment involves recognizing objects, their locations and features such as motion. When a cognitive system has to perceive
multiple objects at once, the binding problem occurs: the system has to extract
█3
features from each entity and ascribe them to the proper item. The problem is
acute with parallel information processing presumably performed in our brain;
how can pieces of information represented in parallel be integrated in a coherent representation? For example, seeing a blue circle and a red rectangle, how
does a brain integrate the representations of blue, red, circle and rectangle as
those of a blue circle and a red rectangle in a particular relative location? In the
current binding problem literature, synchronicity is often mentioned as a possible solution to the problem ([12][3], for example). However, synchronicity alone
does not solve the problem. In the first place, it does not answer the question
of how different objects are represented in the brain. Secondly, considering that
visual scenes are perceived with saccades, there remains a suspicion that visual
perception is rather dynamic than a synchronic process.
An associative system could cope
with the (perceptual) binding problem
in the following manner. Suppose that
the system obtains a feature pattern
of a physical object as well as a pattern representing the coordinates1 of
the object and that it obtains the information of an object at a time. In
this situation, the system shifts ‘ attention ’ among objects to recognize Fig. 1 Association among figure repremore than one object2 . The associa- sentations
tive retrieval of object representation Each figure represents feature patterns for
could be driven by cues representing the figure. The numbers represent cue patorientations (such as 15 degrees left) terns for relative degrees and distances.
and relative distances (Fig. 1). In this
scheme, a scene (containing more than one object) may not be simultaneously
represented, but object representations may be retrieved one-by-one by means
of association in short-term memory or working memory and bound together as
a coherent scene representation. What assures the information binding here is
not synchronicity but the potential or mechanism to bring all relevant representations together.
2.2
Planning
A planning mechanism by means of association could be conceived in a rather
straightforward way as described by the following (pseudo) algorithm:
1
2
Information on the location of physical objects can be obtained via integrating various modalities such as vision, touch, vestibular sense and the sense of eye rotation.
The idea is in line with the Feature Integration Theory [14] but formulated in a more
abstract level.
4
█
if the pattern representing the current situation is not desirable according to
certain criteria, then
associate the current situation pattern (CSP) with an action pattern
(APT), which represents either a directly executable action or a plan.
if CSP and APT are not associated with (do not predict) a pattern
representing a satisfactory situation, then
associate another APT with CSP by suppressing the current APT.
end
if APT represents a learned plan (PL), then
LP: associate APT with a sequence of APTs (APT1, APT2,…).
if the situation pattern associated with (predicted from) CSP and a
sequence consisting of APTi (i=1,2,…) is evaluated as unsatisfactory,
then
abort and suppress PL.
end
if APTi is a learned plan, then
retrieve a sub-sequence for APTi recursively (⇒ LP:).
end
if all the sub-sequences consist of directly executable actions and the
patterns of situations associated (predicted) with the entire APTs and
CSP are satisfactory, then
the entire action (a new plan) is ready to be executed.
end
end
While such an algorithm
is a variation of classical
planning algorithms, some
notes can be given for
that to be realized associatively. In planning, the output sequence could have any
length and plan hierarchy
could have any depth. In
other words, a plan is a cognitive representation having
a combinatory/generative nature. A combinatory representation can be represented
Fig. 2 Planning by means of association
with an associative network,
Ovals are patterns representing the indicated
which actually is a dynamic
content. Blue solid lines represent associative
process of recalling (associrelations.
ating) patterns. In case of a
plan, the entire plan can be represented by retrieving its parts one by one by
means of association. A representation of an action in a plan may be retrieved
5
█
from a representation of the preceding action with a cue pattern representing
the temporal relation succeeding. Here, note that in the sample algorithm above,
a plan is formed (information is bound) with association but without having
recourse to rules with variables.
A plan hierarchy (Fig. 2) would be represented by associating a pattern
representing a learned plan with patterns representing actions or sub-plans3 .
Constructing hierarchical plans requires back-tracking, which, in turn, requires
a mechanism that suppresses an unwanted pattern and retrieves another relevant
pattern (by means of association).
2.3
Language understanding
Understanding language involves syntactic parsing and associating syntactic
structures with semantic representations. As a sentence in a human language
can be indefinitely long, the corresponding associative syntactic/semantic representation can be indefinitely large. Again, such representation has the combinatory/generative nature. While an associative network may not be able to
keep simultaneous activation of patterns for a large representation, the entire
representation could be retrieved piece by piece.
As for semantic representation,
computational models known as
semantic networks (e.g., [11]) are
considered to be associative. A
node of a semantic network represents an individual or type of physical objects, events, situations, time
or location.4 Semantic networks
normally have labeled edges to represent relations between nodes. A
labeled node may be realized in
an associative network as a patFig. 3 A regular phrase structure
tern representing a node and a cue
Blue solid lines can be associative relations,
pattern representing the type of
where ovals represent patterns for syntactic
an edge or relation. For example,
categories.
the edge or relation representing
bigger-than may be realized as association from a pattern representing an individual and a (cue) pattern representing bigger-than to the representation of
another individual.
With regard to parsing, the syntax of human language has recursive structure, i.e., a structure may be embedded in another structure. Such recursive
3
4
[13] can be an exemplary work in this line.
A node in an associative network may accrue meaning by means of non-supervised
learning so that they can be interpreted as a node of a ‘ semantic ’ network. In
fact, a semantic network alone does not give the semantics to its components. See
discussion in [6].
█6
structure would be constructed by a mechanism similar to the one discussed
in the section of planning above, where the construction of sentential structure
would be the goal. The system could traverse a syntactic tree by associating
daughter constituents with the parent node and associating one node to its sibling nodes with cue patterns representing such as left and right (Fig. 3).
Finally, mapping from
syntactic structure to semantic structure must be
taken into consideration.
A pattern representing
a syntactic constituent
such as a clause, verb
and noun may be associated with a pattern representing situation, event
and object respectively.
Here, syntactic relations
may be associated with
Fig. 4 Syntactic and Semantic Patterns
patterns representing semantic relations. For example, an English verb phrase consisting of a verb and
a following noun (having the accusative relation) would be associated with the
representation of an event, an object and the theme relation between them (Fig.
4).
3
Conclusion
The associative representations proposed herein have shared characteristics. In
the three cases, representation is not static but is achieved by dynamically retrieving patterns by association. The associative retrieval is accompanied by cue
patterns such as those indicating spatial and temporal directions and semantic/syntactic relations. The proposed representations also cope with the issues
of dynamicity and generativity and the binding problem. As for dynamicity,
planning and language understanding discussed above are dynamic and the representations proposed here are all dynamic. As for generativity, representations
in planning and language understanding are generative and those in physical
world recognition can also be generative, as a scene is composed of many objects.
The issue of binding without variable was addressed in physical world recognition as well as in language understanding and planning, as these processes must
also bind pieces of information together into coherent representations.
The proposal here apparently requires empirical studies. In particular, if
the illustrated mechanisms are to serve as certain functions, association should
be properly controlled. So, while the author plans to implement experimental
systems for corroborating the proposal, the issue of the (executive) control [8]
shall be seriously addressed.
█7
Acknowledgements
I would like to express my deep gratitude to the comments from the referees of
this paper and also from Quentin Quarles, without which the paper did not form
the current shape.
References
[1] ACT-R: http://act-r.psy.cmu.edu
[2] Fodor, J.A., Pylyshyn, Z.W.: Connectionism and cognitive architecture: A critical
analysis. Cognition 28(1), 3–71 (1988)
[3] Fuster, J.M.: Cortex and mind: Unifying cognition. Oxford University Press (2005)
[4] Gärdenfors, P.: Conceptual spaces: The geometry of thought. MIT press (2004)
[5] Haselager, W., Van Rappard, J.: Connectionism, systematicity, and the frame
problem. Minds and Machines 8(2), 161–179 (1998)
[6] Johnson-Laird, P.N., Herrmann, D.J., Chaffin, R.: Only connections: A critique
of semantic networks. Psychological Bulletin 96(2), 292–315 (1984)
[7] Lakoff, G.: Women, fire, and dangerous things: What categories reveal about the
mind. University Of Chicago Press (1990)
[8] Miyake, A., Shah, P.: Models of working memory: Mechanisms of active maintenance and executive control. Cambridge University Press (1999)
[9] Newell, A.: Unified theories of cognition. Harvard University Press (1994)
[10] OpenCog: http://wiki.opencog.org/w/The Open Cognition Project
[11] Quillian, M.R.: Semantic memory. In: Minsky, M. (ed.) Semantic information
processing, pp. 227–259. MIT Press (1968)
[12] Revonsuo, A.: Binding and the phenomenal unity of consciousness. Consciousness
and cognition 8(2), 173–185 (1999)
[13] Subagdja, B., Tan, A.H.: A self-organizing neural network architecture for intentional planning agents. In: Proceedings of The 8th International Conference on
Autonomous Agents and Multiagent Systems-Volume 2. pp. 1081–1088. International Foundation for Autonomous Agents and Multiagent Systems (2009)
[14] Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cognitive
psychology 12(1), 97–136 (1980)