Basins of Attraction in Fully Asynchronous Discrete

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
397
Basins of Attraction in Fully Asynchronous
Discrete-Time Discrete-State
Dynamic Networks
Jacques M. Bahi, Member, IEEE, and Sylvain Contassot-Vivier, Member, IEEE
Abstract—This paper gives a formulation of the basins of fixed
point states of fully asynchronous discrete-time discrete-state
dynamic networks. That formulation provides two advantages.
The first one is to point out the different behaviors between synchronous and asynchronous modes and the second one is to allow
us to easily deduce an algorithm which determines the behavior of
a network for a given initialization. In the context of this study, we
consider networks of a large number of neurons (or units, processors, etc.), whose dynamic is fully asynchronous with overlapping
updates . We suppose that the neurons take a finite number of
discrete states and that the updating scheme is discrete in time.
We make no hypothesis on the activation functions of the nodes, so
that the dynamic of the network may have multiple cycles and/or
basins. Our results are illustrated on a simple example of a fully
asynchronous Hopfield neural network.
Index Terms—Asynchronism, Hopfield networks, networks dynamic.
I. INTRODUCTION
T
HE role of parallel iterative algorithms is essential in the
domain of scientific computation. An important class of
such algorithms is the discrete-time discrete-state networks
which are useful in numerous applications. Such networks are
usually described as a collection of neurons such that each
neuron takes a finite number of discrete values. If the value
, the global state of the
of neuron is noted ,
and the set of
system is then described by
global states is
where is the finite set of values which can be taken by neuron
. The dynamic of the network is given by the activation function
such that
where each is the activation function of neuron . Those
are supposed to be general and are not restricted to threshold
networks. The global state of the network at the discrete time
(also called iteration ) is denoted by
Manuscript received May 12, 2003; revised July 11, 2005.
The authors are with Laboratoire d’Informatique de L’Université de
Franche-Comté, IUT Belfort-Montbéliard, 90016 Belfort, France (e-mail:
[email protected]; [email protected]).
Digital Object Identifier 10.1109/TNN.2005.863413
Furthermore, the most general execution mode of those networks is the fully asynchronous one. In the literature, different
models of asynchronism are used depending on the way the
communications are managed and the updates are performed
[1]–[3]. In this paper, we consider the fully asynchronous mode
with overlapping updates in the sense defined by Herz and
Marcus in [3].
• The neurons of the network may be updated in a random
order and, moreover, it is possible that some neurons may
not be updated at some times.
• At each time , each neuron updates its own state using
the last received information from the elements (neurons)
it depends on, rather than waiting for their states at time
.
This model corresponds to the most general one which incorporates the sequential, parallel and block-sequential cases. For
a more detailed description of those cases, see [4] and the references therein.
In fully asynchronous networks, since only some of the neurons may be updated at each time , there is a need to define what
, which corresponds to the
is called the strategy, denoted by
set of neurons updated at time . Moreover, since the updating
of each neuron may use values of other neurons computed at
the state of neuron availdifferent times, we denote by
able for neuron at time :
, where
denotes the delay of neuron with respect to neuron . Finally,
some classical conditions are assumed on the
in order to
ensure that the process actually iterates and so evolves.
Definition 1.1: Let us consider an -neuron network and the
, a sequence of subsets of the neurons.
strategy
For
, let
be a sequence of
, such that
c1)
with
,
being the
delay of neuron according to neuron at the discrete
time ;
,
, i.e., although
c2)
the delays associated with neuron are unbounded,
they follow the evolution of the system;
c3) No neuron is neglected by the updating rule. This condition is called fair sampling condition and is equiva,
.
lent to
Then, the fully asynchronous dynamic of the -neuron network associated to the activation function and to the strategy
, and with initial configuration
is described
by Algorithm 1.
1045-9227/$20.00 © 2006 IEEE
398
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
Algorithm 1 Asynchronous iteration
Given an initial state x = (x ; . . . ; x )
for each time step t = 0; 1; . . . do
for each component
if i 2 J (t) then
x
=
f
=
x
x
;.
i
= 1; . . . ; n do
..;x
else
x
end if
end for
end for
Such asynchronism is useful and even sometimes inherent to
natural and physical systems. Moreover, it enables parallel iterative algorithms to be efficiently implemented on a large scale,
using several machines scattered on different sites which can be
far from each other. Such a context is commonly named grid
computing in the literature (see, for example, [5]).
Unfortunately, the dynamic of such asynchronous networks
is far more difficult to estimate and, thus, to control than those
of synchronous ones. Hence, additional studies are necessary
to define the conditions which ensure their convergence. Numerous studies have been done in this domain such as [2],
[6]–[10]. In the same way, a lot of studies have focused on
the global dynamic [7], [11]–[13] or on the conditions to have
cycle-free dynamics in such networks. Nevertheless, to the best
of our knowledge, there is no study which directly deals with
the description of the complete basins of attraction of the fixed
points (see Definition 2.1) of discrete asynchronous networks,
neither in the asynchronism nor in the neural network literature.
This is however an important issue in order to well understand
the behavior of asynchronous networks.
In [4], our goal was to find as many initial conditions as possible which make a network converge toward a given fixed point.
We have expressed sufficient conditions to extend the estimated
basin around a fixed point by using the discrete derivative. We
have then proposed an algorithm which, for a fixed point of the
considered network and an initial input, tells us if this input
makes the network converge to this fixed point in the fully asynchronous mode.
In this paper, we propose a new result which completely describes the attraction basin of a fixed point of a given network in
fully asynchronous mode, whose activation function is known.
An additional result is an algorithm which determines the resulting state of a given network in fully asynchronous mode for
a given input.
According to this scope, the closest works are those of Robert
[1], [14] who studied local attractions in compact neighborhoods of a fixed point (subpart of the basin), and Pellegrin [15]
who proposed some verification algorithms of those attraction
properties. Most of the other studies related to the basins of attraction took place in synchronous networks.
In the particular context of neural networks, some authors
have tried to study asynchronism in this kind of networks [3],
[16]–[19]. Nevertheless, they used a rather quite different approach since most of them focused on conditions on the network
to ensure convergence or cycle absence.
In the context of our paper, we consider general discrete networks which may have several cycles and several fixed points.
The final goal of this paper is to get as much information as
possible about the dynamic of fully asynchronous networks in
order to be able to modify their construction and/or configuration to exactly obtain the desired behavior. The direct application to neural networks is to provide an efficient and accurate
tool to study their dynamic. Hence, we expect that these results
will be helpful in future works dealing with the enhancement
of the design of recurrent neural networks used, for example, as
associative memories.
In Section II, all the notions and definitions required for the
description of the attraction basin of a fixed point are presented
together with the underlying mechanisms and justifications.
Then, the definition of the attraction basin is given in Section III
together with a deduced formulation of the global dynamic of
an asynchronous network. The proof of this result is placed
in Appendix I for a faster reading. The mechanisms of the
construction of the basin are pointed out in some examples in
Section IV. Then, a brief discussion about the generalization of
our result to block-decomposed networks is given in Section V
and a determination algorithm based on our formulation is described in Section VI. An example of application to a Hopfield
network is finally given in Section VII.
II. PRELIMINARY DEFINITIONS
A. Generalities
Definition 2.1: A fixed point state
verifies
which implies in the context of Algorithm 1 that
,
Once such a state has been reached, the network remains in
this state forever although its components are updated. In other
words, the updating of each component leads to the same value.
In this case, we say that the network has converged to this fixed
point state. That kind of state must be distinguished from those
of stable state. The latter is an extension of the first one in which
an additional constraint is set on the neighborhood of the state.
Hence, a stable state is a fixed point state whose neighboring
states (at least all the ones which have one different component
from it) also lead to it in a finite number of iterations. It is important to notice that this is not necessary the case for a fixed
point state. In the remaining of the paper, we use the expression
fixed point as a shortened form for fixed point state.
is
Definition 2.2: The basin of attraction of a fixed point
in a
defined as the set of all the states which surely lead to
finite number of iterations.
Hence, all vectors belonging to a cycle in the iteration graph
or which can lead to different fixed points from one execution to
another cannot be included in such a basin. To study the basins
of fixed points in such asynchronous systems, we use the discrete distance between two states and of the network given
by
(1)
BAHI AND CONTASSOT-VIVIER: BASINS OF ATTRACTION IN DISCRETE-TIME DISCRETE-STATE DYNAMIC NETWORKS
where
if
if
.
This so-called vectorial distance was introduced by Robert [1],
[14] in the context of component-wise chaotic discrete boolean
iterations, which is a particular case of block-asynchronous iterations on finite sets, and extended to finite sets in [20].
In the following, we suppose that we know the activation
of the network, and we want
function and a fixed point
to find all the vectors which surely make this network converge
in fully asynchronous mode.
to
399
In the first case, all the components will have the same value
at the next iteration independently of the updates. In the second
case, there is only one component which may change. In this
case, we are still confronted to indeterminism since there are
two possibilities (the current state or its image by ). Nevertheless, the hypothesis made on the delays in fully asynchronous
mode ensures us that updates occur in a finite time. Hence, the
evolution of the system from may be itself during a finite
. This means that those two
time and then will surely be
cases can be assimilated to deterministic evolutions.
In the following part, we give some useful tools required for
the formulation of the basin of a fixed point.
C. Additional Tools
B. Fundamental Difference Between Synchronous and
Asynchronous Iterations
All the difference between the synchronous and asynchronous modes lies in the number of possible successors of a
given state of the system.
In the synchronous case, the process is completely deterministic and, to each state of the system, there corresponds only one
successor.
This is not the case in the asynchronous mode. To each state
of the system, there may correspond several possibilities of evolution for the next iteration depending on which components are
updated.
and
Let us consider the example of a 3-node system in
. Then, in syna transition function such that
chronous mode, the state following 011 is surely 000 whereas
there are four possible outcomes in the asynchronous case depending on which components are updated. In our example, the
first component stays the same whether it is updated or not.
Thus, the possible evolutions only depend on the updates of the
last two components. In the following list of the four possible
cases, we do not specify the behavior of the first component
since it leads to the same value:
• 000: The last two components are updated ( sync case).
• 001: The third component is not updated.
• 010: The second component is not updated.
• 011: The last two components are not updated.
Hence, we clearly see here all the difference between those
two kinds of iterations. Synchronous iterations are deterministic
whereas asynchronism induces a nondeterministic evolution of
the system.
It is important to notice that there may be some components
for which the updating is equivalent to the standby. These components are the ones whose values do not change between and
. In our example, it is the case of the first component whose
evolution at the following iteration is the same for updating and
standby. This allows us to give the local conditions under which
the asynchronous mode of execution is equivalent to a deterministic case.
For a given state of the system, we can deduce two cases
where asynchronous iterations are equivalent to synchronous
ones:
• the state is a fixed point;
.
• there is only one different component between and
For any couple of vectors
, we define the set of vec, which contains all the possible mixings of
tors
the component values of and . This construction directly depends on the nonzero elements in the vectorial distance
.
Definition 2.3:
where
if
if
.
We also define the asynchronous successor relation between
), which means that
two vectors and (denoted by
and can be two consecutive states in the asynchronous evolution of the system.
,
, if and only if
Definition 2.4:
and we extend that notion to the asynchronous iteration path
relation between vectors and (denoted by
), which
means that state can lead to state through a sequence of
asynchronous iterations.
,
if and only if there exists a
Definition 2.5:
,
, possibly empty,
sequence
of distinct states of such that
To describe the basin of a fixed point, we also need to define
what we consider to be a cycle in a fully asynchronous network.
D. Characterization of Cycles and Pseudocycles
The difficulty to characterize the cycles in the fully asynchronous mode comes from the fact that, contrary to the
synchronous mode, there are two possible kinds of cycles
in the asynchronous case: The cycles which are attractors of
the system and the ones which are not. Obviously, only the
attractor cycles are relevant in the study of the dynamic of
such networks. Those cycles are described in this subsection
together with complementary notions.
In the asynchronous mode, the general evolution of the
system is directly linked to what in the literature are commonly
called pseudoperiods [21].
400
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
Fig. 1. Pseudoperiod.
Definition 2.6: A sequence of states
,
is a pseudoperiod of a fully asynchronous network associated to
the set of states and the activation function , if and only if
Fig. 2. Pseudocycle does not contain a complete pseudoperiod.
and
Fig. 3.
where denotes the component-wise logical
operator.
and folThis definition means that starting from the state
, the state
is the first state
lowing the sequence
at which all the components of the system have been updated at
least once since . The first condition ensures that there is no
component which is not updated in the sequence. If the compois updated at the following iteration, then
nent of
we have
and
. Thus, to obtain a null vector, all the components must be updated at least
is
once in the sequence. The second condition ensures that
necessary to have a null vector and, thus, is the first state in the
sequence at which all the components have been updated at least
once.
The necessity of definition 2.6 comes from the fact that, on
the one hand, Algorithm 1 implies that, at each iteration, some
components of the system may not be updated (inducing indeterminism in the system). On the other hand, according to hypotheses c3) in Definition 1.1, it is not possible to have a component which is never updated. Thus, for any time during the
such that all the
execution, there exists a finite integer value
components of the vector have been updated at least once be. Fig. 1 gives an example of a pseutween iterations and
doperiod in a system of four components beginning in A and
finishing in B. The arrows represent the iterations and the arrays
indicate which components of the system are updated at each iteration (white cells not updated, black cells updated). It can
be seen in this figure that the superimposition of all the arrays
gives an array full of black cells. Moreover, some components
are updated several times in this sequence.
Thus, concerning the cycles in such networks, there are two
possibilities based on the fact that either the cycle does not contain any complete pseudoperiod or contains at least one.
The first possibility corresponds to what we call pseudocycles. This naming comes from the fact that, according to the
asynchronous model, it is not possible to stay infinitely in such
a cycle. Let us consider the example given in Fig. 2 and suppose that the system enters the cycle through the state . It can
be seen that when the system goes through the cycle and comes
back to , the third component has not been updated. This and
the fact that there is no complete pseudoperiod in the sequence
imply that the updating of that third component for any state
in the cycle will lead to another state outside that cycle. This
Cycle contains at least one complete pseudoperiod.
leaving of the cycle will happen in a finite time. Thus, this cycle
cannot be gone through infinitely and then it does not represent
the attractor of the system.
By symmetry, the second possibility corresponds to the attractor cycles which are simply called cycles. Since they contain
at least one complete pseudoperiod, as can be seen in Fig. 3,
there is no condition which forces the system to leave such a
cycle which can thus be gone through infinitely. Thus, they can
be considered as attractors of the network.
Finally, the cycles of an asynchronous network can be defined
as follows.
Definition 2.7: Let us consider a fully asynchronous network
associated to the set of states and the activation function .
,
of states of forms
The sequence
a cycle of length in the asynchronous iteration graph of this
network, if and only if
and the sequence
contains at least one pseudoperiod.
It can be noticed that this formulation also includes the fixed
points of the system since they form a cycle of length one in the
.
particular case where
the set of states
In the same way, we denote by Cycle
belonging to at least one cycle in the asynchronous dynamic of
the network associated to the set of states and to the activation
function . Similarly, the set of states in pseudocycles is given
in the following definition.
Definition 2.8: The set of states outside any cycle and involved in at least one pseudocycle containing a given vector
Cycle
is defined by
Cycle
such that
Cycle
This description specifies that there exists an iteration path
from state to itself which goes through state . However,
since all the elements of the sequence
are taken outside
BAHI AND CONTASSOT-VIVIER: BASINS OF ATTRACTION IN DISCRETE-TIME DISCRETE-STATE DYNAMIC NETWORKS
Cycle
, this path can only be a pseudocycle. It can be
Cycle
,
since,
noticed that
.
by definition, we always have
Finally, those previous definitions allow us to describe the
asynchronous convergence toward a fixed point (denoted by
).
be a fixed point of the network assoDefinition 2.9: Let
ciated to the set of states and the activation function .
,
, if and only if
and
such that
and
Cycle
Using all those definitions, we can now describe the basin of
a fixed point state in a fully asynchronous network.
III. BASIN OF A FIXED POINT
(denoted by
Definition 3.1: The basin of a fixed point
) of a fully asynchronous network associated to the set
of states and the activation function can be described by the
,
, recursively defined as
union of the sets
follows:
Cycle
and
let
401
mutual dependence for the inclusion. We can do that because
all the elements of a pseudocycle are equivalent in terms of evolution (this can be easily verified). Thus, not taking into account
the other elements of a pseudocycle does not make us miss any
vector of the basin.
is
Finally, the last important remark is that
. Effectively, although the
always smaller than
vector
is in
, it cannot be in
.
The contrary would imply the presence of a pseudoperiod in an
to
through and, thus, a cycle,
iteration path from
leading to a contradiction with Definition 2.8. Hence, the set
is always nonempty. This definition leads to the following
description of the dynamic of fully asynchronous networks.
Theorem 3.2: Let be a fixed point of a network associated
to the set of states and the activation function . If we consider
, then we have the following.
, all the asynchronous executions of the
•
system starting from lead to the state .
, asynchronous executions starting from
•
may either:
, meaning
— always lead to the same fixed point
all the executions of the network will lead to the same
fixed point different from ;
— lead to different fixed points (eventually to
),
meaning different executions of the network will lead
to different fixed points;
— lead to a cycle, meaning executions of the network may
not lead to a fixed point and so may not terminate.
For the complete proof of this theorem, see Appendix I.
IV. SOME EXAMPLES
In order to exhibit the construction mechanism of the
sets, several examples corresponding to different cases are presented.
The goal of our first example is to clearly show the increset, and to point out the
mental way of construction of the
particular order of inclusion of the elements in its subsets. It
consists of a particular case of global convergence toward the
. The graphical representation of its tranunique fixed point
sition function is given in Fig. 4.
From , we obtain the following sets:
Cycle
This description takes into account all the indeterminism induced by the full asynchronism. It is based on the fact that one
in the fully asynchronous mode if
state will always lead to
and only if all the possible evolutions starting from lead to
without going through any cycle. Hence, the basin is recursively
built by using a reformulation of this convergence property: A
state is in the basin of in the fully asynchronous mode if and
only if all its possible successors at the next iteration are also in
that basin. This is why the description in Definition 3.1 is based
, which allows us to describe the whole set of
on
possible asynchronous successors of at the next iteration.
Moreover, cycles and pseudocycles are also taken into account. All the elements in cycles are systematically discarded
and elements in pseudocycles are not directly involved in the
inclusion of a given vector since they would clearly lead to a
Then, the basin of 000 is fully described by the following
subsets:
402
Fig. 4. Graphical representation of transition function g .
In this example, it can clearly be seen that each element is
subset only if all its possible asynincluded in an
chronous evolutions at the following iteration, except those
subin pseudocycles, are already in the previous
which is
sets. This is particularly striking for the vector
the last one included in the set whereas it is closer to
than some other vectors and its transition is directly
.
In fact, it can be included only when all the elements in
are
already included. The second element of this set being only in,
cannot be included before the
cluded in
subset.
There is no cycle in this example nor in the following one. Yet,
such a case is given in the last example. Moreover, in the folset as the union
lowing examples, we directly write each
subsets in their building order without naming
of their
them.
Another example is the possibility for the network to converge toward different fixed points in a series of executions when
initialized with a same vector. This case is assimilated to a divergence and can be obtained by a slight modification of the
and
transition function into a function so that
for the other vectors. Its graphical representation
is given in Fig. 5.
Then, the two basins are
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
Fig. 5. Graphical representation of function h.
Fig. 6. Graphical representation of function w .
included in a unique basin. Here again, our construction given in
Definition 3.1 does not include such a point in any of the basins.
The last example which may arise is the presence of a cycle in
the asynchronous iteration graph. If we use the transition function , a graphical representation of which is given in Fig. 6,
and
form a cycle in the
it can be verified that vectors
asynchronous mode.
Nevertheless, in order to exhibit the mechanisms of Definition
.
2.7, we detail the deduction of the cycle
implies
As seen in previous examples,
Thus,
. Moreover, since
and the following chain can be built:
, we have
and
and
and
is not included in any of the
sets since it may lead
or to
from an execution to another. This comes
either to
, since
from the fact that, when starting from
, as already detailed in Section II-B with another example,
there are four possible evolutions (
and
) in
fully asynchronous mode, depending on the eventual updating
of each of the last two components.
, as
The first and third cases lead to convergence toward
already seen. The second case can only evolve toward
since
and
.
there is only one difference between
. The fourth case is the initial vector itself and,
It leads to
according to the hypothesis on the delays, it is sure there will be
a time at which at least one of the last two components will be
updated, leading to one of the other cases.
may
Thus, it can be seen that executions starting from
or to
implying that this vector cannot be
lead either to
This means that
quence (
is actually a cycle since the se) contains a pseudoperiod, and we have
Cycle
Another interesting aspect of this example is the pseudocycle
and
. Effectively, if we consider their
formed by vectors
respective
sets, we have
and
BAHI AND CONTASSOT-VIVIER: BASINS OF ATTRACTION IN DISCRETE-TIME DISCRETE-STATE DYNAMIC NETWORKS
We can build the chain
which denotes a potential cycle. Nonetheless, the evaluation of
its corresponding vector gives
403
where each ,
, is the collection of the activation
functions of the neurons in block .
Since each
is a collection of finite-state components, its
set of values, , is also a finite set of discrete values given by
and the set of global states can then be reformulated as
which reveals that this chain does not contain a pseudoperiod.
) is a pseudocycle, and we have
Thus, the sequence (
Hence,
and
are included in
since all their possible evolutions except those in their pseudocycles are either in
or in
.
The results for this example are then
and
Finally, we can see that Definition 3.1 allows us to build the
basins of all the fixed points of a fully asynchronous network
whose transition function is given. In Section V, we explain how
this result can also be applied to block-decomposed networks.
V. DISCUSSION ABOUT BLOCK-DECOMPOSED NETWORKS
To be clearer, the results given in this paper have been presented in the context of a simple discrete-state discrete-time network, in which each component evolves asynchronously according to the other components. However, those results also
hold in the context of block-decomposed networks. This comes
from the fact that a block-decomposed network can be reformulated as a nondecomposed discrete-state discrete-time network
as shown in this section.
Starting from the formulation of an -neuron network used
in the previous sections, we have
where each
is the finite set of the values which can be taken
by neuron .
blocks, each one
Let us partition this network into
neurons. Thus, the blocks
,
containing
are distinct nonempty subsets of the set of components
. The global value of the network is described
. In this context, it can be noticed that
by
is a particular case of
where
and each block contains exactly one neuron.
Concerning the dynamic of the block-decomposed network,
the activation function is partitioned in a compatible way into
a function such that
Thus, for a network with components ,
, associated to the set of states and to the activation function , a
block-decomposition of this network into blocks can be seen
,
as a nondecomposed network with components ,
and to the activation function
associated to the set of states
. In this way, the results presented in this paper can also be
applied to block-decomposed networks.
In Section VI, we propose an algorithm, based on our formulation of the basins, which determines, for a given initial vector,
the result of the fully asynchronous evolution of a given network.
VI. DETERMINATION ALGORITHM
In the first part of this section, we propose an algorithm which
determines, for an initial state , the result of the fully asynchronous iterations of a network whose activation function is
given. The second part is dedicated to the evaluation of the complexity of that algorithm.
A. Algorithm
As explained in [4], since the synchronous mode is a particular case of the asynchronous one, a necessary condition to have
starting from
in fully
convergence toward the fixed point
asynchronous mode is that leads to in synchronous mode.
Hence, our testing algorithm is decomposed into two main steps
reached when starting from
•
find the fixed point
in synchronous mode;
•
if it exists, test if
is in the basin of
in asynchronous mode.
In the first step, cycles may be detected in the synchronous iteration graph. If this is the case, the second step is not necessary
the system may not
since we can directly deduce that from
converge in a finite time in fully asynchronous mode. In fact, the
nondeterministic behavior of asynchronism may open the cycles
in the iteration graph. This means that if the network is in a state
belonging to a cycle, there may be an evolution which will lead
to a state outside the cycle. If so, there is a possibility to reach a
fixed point. Nevertheless, this cannot be seen as a convergence
since the cycle in the iteration graph still exists and the leaving
of the cycle is not ensured in a finite time.
Algorithm 2 details the first step of the process. It consists
in parsing the iteration graph of the synchronous mode starting
from in order to find either a cycle or a fixed point. To manage
cycle detection, a history of the transient states in the traversal of
404
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
the iteration graph is stored in the stack stateList. When a fixed
point is found, Algorithm 3 is executed.
= x is already in stateList, we have to test if
the chain of vectors from the top of stateList down
to the first occurrence of x is a true cycle =
;
x
x
6
while x = f (x) and x
add x to stateList
f
)
y
62 stateList do
P
y
x
;
else
in sync mode
!
=
This last algorithm is formulated as a recursive function
which parses the iteration graph of the asynchronous mode
starting from . Its parameters are the current state in the
recursion, the fixed point of the eventual convergence, a stack
stateList containing all the states taken by the system to arrive
at the current state from the initial state , and the activation
function of the network . As in Algorithm 2, this stack is
necessary to detect cycles in the iteration graph and is empty at
the first call of the function.
At each recursive call, we make several tests on the current
state . According to Definition 3.1, we have to make the distinction between a state which is in the basin of , a state which
may lead to other fixed points and a state which belongs to a
cycle. All these cases are addressed in Algorithm 3.
Algorithm 3 function testAsync(x, x , stateList,
if x = x then ==x leads to x in async mode
return True
f
)
else
if x = f (x) then = x can lead to another fixed point
than x in async mode =
return False
True
==
==
test of all evolutions
init of the parsing result
add x to stateList = push x in the stack for the
recursive parsing of the iteration sub-graph =
for all z C P (x; f (x)) x do
res AND testAsync(z , x , stateList, f )
res
2
AND
==
d(f (z ); y )
get the previous state in the cycle
(0; . . . ; 0) then
==
a true cycle has been
returning True here, allows us to continue the for
all loop above to test the other branches (at least
one) =
return True
x
62 stateList then
x
found
return False
else = there is only a pseudo-cycle in this
branch, we cannot state the final result yet, but
=
is already in stateList
cycle detection
does not lead to convergence
end if
= x
P
z
end for
=
if P
end while
if x = f (x) then
= x
leads to the fixed point
testAsync(x , x, , f )
else
if x
from x
res
x==
do
f (x )
x
(1; . . . ; 1)
== init of the boolean product
init with the last state of the cycle
for each vector z from top of stateList downto
P
Algorithm 2 function testSync(x ,
stateList
nf g
if res = False then no need to continue
Stop the for loop
end if
end for
remove x from stateList = pop x from the stack
to come back at the current parsing level =
return res == result for the sub-graph from x
else
end if
end if
end if
end if
The first condition corresponds to the terminal case of a
branch of the iteration graph which leads to . All the recursive calls of the algorithm will arrive in this case for a
state which actually belongs to the basin of . According to
subset of
Definition 3.1, this case corresponds to the
.
The second condition corresponds to the divergences since
whereas
. In this case, convergence toward a
unique fixed point is not assumed, and then cannot belong to
the basin of .
The third case is the general recursion on the iteration graph
. If the current state is not
to determine if belongs to
in stateList, it means that it has not been traversed yet and all
asynchronous iterations starting from must be tested. This is
equivalent to evaluate the subpart of the asynchronous iteration
graph whose root is . This is recursively performed for every
possible evolution of and the results of all the branches are
aggregated to get the final evaluation. It is enough for one branch
not to lead to in order to detect that is not in the basin of .
The recursion in this determination algorithm proceeds in the
set described
opposite way of the construction of the
set is built from inner
in Definition 3.1. Indeed, although the
to outer
subsets, the determination algorithm can be seen as
a parsing of the subsets from outer to inner ones. Hence, for a
given vector , if all the elements in
converge to , it means that there exists a finite such that those
elements belong to
. Then, we can deduce from
and, thus, to
.
Definition 3.1 that belongs to
Finally, the last case corresponds to the possibility that the
current state is in a cycle. If the current state is already in the
stateList stack, it means that it has already been traversed and we
have to test if the chain of states from the top of stateList down
to the first occurrence of is a true cycle or just a pseudocycle.
When a cycle is detected, the convergence cannot be guaranteed
BAHI AND CONTASSOT-VIVIER: BASINS OF ATTRACTION IN DISCRETE-TIME DISCRETE-STATE DYNAMIC NETWORKS
in a finite time and the determination process is stopped. When
a pseudocycle is detected, this branch of the iteration graph does
not pose any problem for the convergence. In fact, the final result
of the determination algorithm will be obtained by the parsing
of the other branches of the iteration graph outside the pseudocycle. Previous discussions about pseudocycles have pointed
out that such other branches exist.
That description of our determination algorithm clearly
shows its direct correspondence to the description of the
set given in Definition 3.1.
B. Complexity
The complexity of our determination algorithm is directly
linked to the complexity of the traversals of the iteration graphs
of the synchronous and asynchronous modes. It can be quite
large in the worst case since each traversal is proportional to
the number of states of the system and, thus, it is exponential
in function of the number of nodes. Nonetheless, the complete
traversal of the graph is a rare case in practice since it can only
occur in contractions, where all the states lead to a unique fixed
point. In most cases, there will be several fixed points and, thus,
several basins, each of them smaller than the whole set of states.
As our algorithms are designed to traverse a minimal number of
states necessary to perform the determination and since, in most
cases, this number is far less than the total number of states, we
. Moreover,
obtain an average complexity far less than
this also holds for the nonconverging cases since the algorithms
are designed to stop as soon as they find a branch of the iteration
graph leading to a cycle or to a divergence.
In the following, we give the complexity of the determination
algorithm for a given network whose set of states , activation
are given. Since the behavior of
function and initial state
the algorithm is not the same for the three possible kinds of results (convergence, cycle, and divergence), we give the complexity for each case.
1)
Convergence: In case of convergence toward a fixed
point , the complexity is proportional to the number
of states in the basin containing
Complexity
and more precisely, if we know that
then
Complexity
2)
Divergence: In this case, there exist at least two fixed
points
and
which can be reached in different
asynchronous executions initialized by the same . In
this context, it is sure that one of those fixed points, say
, is the result of the synchronous executions starting
from . Thus, we have
Complexity
where
denotes the length of the synto
, and
chronous iteration path going from
3)
405
is the length of an asynchronous iteration
to
.
path from
Cycle: In this last case, there are two possibilities of
detection depending on whether the cycle is in the
synchronous or in the asynchronous iteration graph. In
the first case, the detection is performed by Algorithm
2 whereas in the second case, it is performed by Algorithm 3. This implies two different complexities which
are given below.
If the cycle is in the synchronous iteration graph,
we have
Complexity
where is the first state of reached from , and
denotes the length of cycle .
If is in the asynchronous iteration graph, we obtain
Complexity
Concerning the space complexity, it is directly related to the
length of the stack stateList which contains the current path in
the iteration graph. Hence, it depends on the number of traversed
states and, thus, has a similar expression as the time complexity
for each of the previous cases. The only difference is an additional factor which is the memory needed to store one state ,
which is itself directly related to the dimension of (i.e., the
number of nodes in the system).
VII. APPLICATION
In this section, we take the same example of application as in
[4] in order to be able to compare the results of our algorithm
presented in the previous section to the one presented in that
previous paper. Our previous algorithm was based upon local
features and especially the discrete derivative of the activation
function. This previous work was based on the contraction property of the local derivative while the technical framework of the
present work is based on nested sets.
This example takes place in the context of Hopfield networks
and the reader should refer to [4] to have a detailed state of the
art on Hopfield networks and their connections to discrete-time
discrete-state fully asynchronous networks.
In the following, a Hopfield network whose value of each
, is built from a given set of vectors to
neuron is in
be memorized (see Table I) using the Hebb’s rule. Then, with
the initial vectors given in Table II (noisy patterns), the convergence of the asynchronous network is tested using the algorithm of [4] and the version given in Section VI. The last
column corresponds to the actual result of the network obtained
, whose value is
by a simulation algorithm. The fixed point
, corresponds to a spurious state
induced by the building method of the Hopfield network.
As expected, our algorithm presented in Section VI gives correct results to all tested vectors, which was not the case for our
previous version presented in [4].
In terms of complexity, the number of traversed states is given
for each tested vector in Table III.
It can be seen that the actual number of traversed states is far
. This
less than the total number of states which is
406
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
SET
OF
TABLE I
MEMORIES (DESIRED FIXED POINTS)
1 AND
1)
(
0
TABLE II
TESTS OF CONVERGENCE FOR EIGHT VECTORS (
OF THE
NETWORK
1 AND 01)
The applicative field has not been neglected since an algorithm has been deduced from this formulation, which gives the
asynchronous behavior of a given network for a given input
state. When the network converges, the algorithm identifies the
resulting fixed point. Otherwise, the algorithm determinates if
the nonconvergence of the network is due to a divergence or to
a cycle.
Finally, this determination algorithm allows us to better study
the behavior of fully asynchronous networks. It can then be
useful to study the dynamic of asynchronous neural networks
and the possible enhancements to their building. Our next goal
is to find rules which should be used to obtain more efficient and
accurate neural networks.
APPENDIX
PROOF OF THEOREM 3.2
TABLE III
NUMBER OF TRAVERSED STATES DURING THE DETERMINATION
EIGHT TESTED VECTORS
OF THE
has a direct impact on the performances of the algorithm and, in
fact, all the determinations presented here have been performed
in real time on a classical PC machine.
Finally, by allowing us to know the complete behavior of
this kind of asynchronous network, our determination algorithm
provides a very accurate tool to study the possible generic optimizations for the construction of neural networks. This objective is of essential interest since the methods used to build neural
networks are often problem dependent. We can reasonably expect that our determination algorithm will help to design a new
kind of generic construction algorithms of neural networks.
The different cases listed in the second part of Theorem 3.2
are not directly detailed in the following proof since they are, in
fact, direct deductions of all the other possible results of such
. They are
a network for any initial vector outside of
listed in Theorem 3.2 more for completeness and accuracy than
as an important part of the theoretical result itself.
Hence, the proof of Theorem 3.2 is made in two steps
,
.
A)
B)
,
and its
Case A) In this case, we consider a fixed point
, and show that this set:
corresponding set
1)
does not contain any other fixed point than ;
2)
does not contain any element of a cycle other than ;
3)
. Obviously, the
Case A.1) Let us consider
implies that
definition of Cycle
Cycle
. This prevents to be in any of
subsets for
according to Definithe
tion 3.1. Hence, the only possibility for to
is in the
subset, which
be in
implies that it is
itself; so,
is the only
.
fixed point in
Case A.2) This case is a direct deduction from Definition
3.1 and Case A.1).
. This implies that
Case A.3) Let
such that
VIII. CONCLUSION
A theoretical formulation of the basins of fixed points of
fully asynchronous discrete-time discrete-state networks has
been given. This formulation directly implies the possibility to
describe all the basins of such networks and then to completely
know their behavior. A formulation of the dynamic of such
networks has then been deduced and its correctness has been
proved. This work has also induced the characterizations of
cycles and pseudocycles in such asynchronous networks which
also represent interesting theoretical results. Moreover, a justification of the validity of those results on block-decomposed
networks has also been given.
If
,
which directly leads to
.
For
, Definition 3.1 implies that
and
As discussed in Section III, the set in Definition 3.1 cannot be empty, which ensures
BAHI AND CONTASSOT-VIVIER: BASINS OF ATTRACTION IN DISCRETE-TIME DISCRETE-STATE DYNAMIC NETWORKS
the existence of at least one vector . Moresubsets, it is
over, by construction of the
desure that there is at least one of the
scribed above which is in
. In the
opposite case, would belong to one of the
,
, instead of
. To
subsets
be simpler, let us only consider the element
, which implies, in turn, that
and
In addition, by recursion on the
we obtain
subsets,
implying, for all the vectors
,
The last line, according to Definition 3.1, leads to
, which is in contradiction with
.
Thus, we have
.
ACKNOWLEDGMENT
and
which ends with
for
obtain a sequence of vectors
, such that
. Thus, we
,
which implies
.
and
Case B) In this case, we consider a vector
we show that it is not possible to have
.
such that
Let us suppose that
and
. According to the transition
function , we have two possibilities
In the first case,
together with the hypothimply that
, which is a contraesis
diction with
. In the second case, from
, we can directly deduce that
the definition of
and then
leading to the recursion
which ends for some
407
, when
The authors would like to thank Mrs. A. Borel for her linguistic contribution.
REFERENCES
[1] F. Robert, Discrete Iterations, A Metric Study. Berlin, Germany:
Springer-Verlag Series in Computational Mathematics, 1986, vol. 6.
[2] D. Bertsekas and J. Tsitsiklis, Parallel and Distributed Computation. Englewood Cliffs, NJ: Prentice Hall, 1999.
[3] A. Herz and C. Marcus, “Distributed dynamics in neural networks,”
Phys. Rev. E, vol. 47, no. 3, pp. 2155–2161, 1993.
[4] J. Bahi and S. Contassot-Vivier, “Stability of fully asynchronous discrete-time discrete-state dynamic networks,” IEEE Trans. Neural Netw.,
vol. 13, no. 6, pp. 1353–1363, Nov. 2002.
[5] Grid Computing Info Centre [Online]. Available: http://www.gridcomputing.com/
[6] J. Bahi and C. Michel, “Convergence of discrete asynchronous iterations,” Int. J. Comput. Math., vol. 74, pp. 113–125, 2000.
[7] M. E. Tarazi, “Some convergence results for asynchronous algorithms,”
Numer. Math., vol. 39, pp. 325–340, 1982.
[8] P. Koiran, “Dynamics of discrete-time, continuous-state Hopfield networks,” Neural Comput., vol. 6, pp. 459–468, 1994.
[9] L. Wang, “On the dynamics of discrete-time, continuous-state Hopfield
neural networks,” IEEE Trans. Circuits Syst. II, Analog Digit. Signal
Process., vol. 45, no. 6, pp. 747–749, Jun. 1998.
[10] X. Wang, A. Jagota, F. Botelho, and M. Garzon, “Absence of cycles in
symmetric neural networks,” Neural Comput., vol. 10, pp. 1235–1249,
1998.
[11] D. Chazan and W. Miranker, “Chaotic relaxation,” Linear Algebra Appl.,
vol. 2, pp. 199–222, 1969.
[12] G. Baudet, “Asynchronous iterative methods for multiprocessors,” J.
Assoc. Comput. Mach., vol. 25, pp. 226–244, 1978.
[13] V. Kozyakin, A. Bhaya, and E. Kaszkurewicz, “A global asymptotic stability result for a class of totally asynchronous discrete nonlinear systems,” Math. Control, Signals Syst., vol. 12, no. 2, pp. 143–166, 1999.
[14] F. Robert, Les Systèmes Dynamiques Discrets. Berlin, Germany:
Springer-Verlag, 1995, vol. 19.
[15] D. Pellegrin, “Algorithmique Discrète et Réseaux D’automates,” Ph.D.
dissertation, Grenoble, France, 1986.
[16] E. Golès, F. Fogelman-Soulie, and D. Pellegrin, “Decreasing energy
functions as a tool for studying threshold networks,” Disc. Appl. Math.,
vol. 12, pp. 261–277, 1985.
[17] C. Marcus and R. Westervelt, “Dynamics of iterated-map neural networks,” Phys. Rev. A, vol. 40, no. 1, pp. 501–504, 1989.
[18] A. Bhaya, E. Kaszkurewicz, and V. Kozyakin, “Existence and stability of
a unique equilibrium in continuous-valued discrete-time asynchronous
Hopfield neural networks,” IEEE Trans. Neural Netw., vol. 7, no. 3, pp.
620–628, May 1996.
[19] C. Marcus and R. Westervelt, “Dynamics of analog neural networks with
time delay,” in Advances in Neural Information Processing Systems I, D.
Touretzky, Ed. San Mateo, CA: Morgan Kauffman, 1989.
[20] J. Bahi, “Boolean totally asynchronous iterations,” Int. J. Math. Algorithms, vol. 1, pp. 331–346, 2000.
408
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 17, NO. 2, MARCH 2006
[21] J.-C. Miellou, “Algorithmes de relaxation chaotique à retard,” Revue
Française d’Automatique, Informatique, Recherche Operationnelle,
Série Analyse Numérique (RAIRO, R-1), pp. 52–82, 1975.
Jacques M. Bahi (M’03) received the Ph.D. degree
and the HDR degree in applied mathematics at the
University of Franche-Comté, Belfort, France, in
1998.
He is a Full Professor of Computer Science at the
University of Franche-Comté. His research interests
include asynchronism, distributed computing, load
balancing, and massive parallelism.
Dr. Bahi is a member of the IEEE Computer Society.
Sylvain Contassot-Vivier (M’03) received the
Ph.D. degree from École Normale Supérieure, Lyon,
France, in 1998.
He obtained a postdoctoral position at the Facultés
Universitaires Notre-Dame de la Paix (FUNDP),
Namur, Belgium, in 1998. Since 1999, he has been
an Assistant Professor at the Computer Science Laboratory (LIFC) at the University of Franche-Comté,
Belfort, France. His main research interests include
massively parallel systems, asynchronism, grid-computing, image processing, and vision.
Dr. Contassot-Vivier is a member of the IEEE Computer Society.