Stable Dynamic Backpropagation Using Constrained Learning Rate

Proceedings o f I993 International Joint Conference on Neural Networks
Stable Dynamic Backpropagation Using Constrained
Learning Rate Algorithm
Liang Jin, Madan M. Gupta, Peter N. Nikiforuk
Intelligent Systems Research Laboratory
College of Engineering, University of Saskatchewan
Saskatoon, Saskatchewan, Canada S7N OW0
ABSTRACT
Equilibrium point learning problem in discrete-time dynamic neural network is studied in this paper
using stable dynamic propagation with constrained learning rate algorithm. The new learning scheme
provides an adaptive updating process of the synaptic weights of the network, so that the target pattern
is stored at a stable equilibrium point. The applicability of the approach presented is illustrated through
a binary pattern storage example.
1. INTRODUCTION
Dynamic back propagation (DBP) concept was first proposed by Pineda [l,21 for the supervised
weight learning problem of a class of recurrent neural networks. For equilibrium point learning problems, however, D B P algorithm provides an effective weight learning process such that the target pattern
is directly stored at an equilibrium point of the network, the stability of the network in the sense of
Lyapunov is not guaranteed at each iterative instant, and the computational requirments increase due
t o the instability. In this paper, a constrained learning rate algorithm for DBP is introduced to avoid
the instability of the network in the dynamic weight learning phase.
2. DYNAMIC NEURAL NETWORKS
2.1 Network Models
Consider a general form of dynamic recurrent neural network described by a discrete-time nonlinear
system
E@
1) = f [ x ( k ) w,
, SI
(1)
Y(k)
= h[x(k)l
where E = [q,. ,E"] E R" is the state vector of the dynamic neural network, z; represents the
internal state of the i - t h neuron, W = [ ~ ; , j ] n xisn the real-valued matrix of the synaptic connection
weights, s E R" is an input vector, y = [ y l , . . .,ymIT is an observation vector or output vector, f :
R" x Rnxnx R"
R" is a continuous and differentiable vector valued neural activation function,
f;(.) and af;/as are respectively bounded and uniformly bounded, and h(5): R" 4Rm is a known
continuous and differentiable vector valued function.
The recurrent neural network consists of both feedforward and feedback connections between the
layers and neurons forming complicated dynamics. In fact, the weight wij represents a synaptic
connection parameter between the i - t h neuron and the j - t h neuron. Hence, the nonlinear vector
valued function on the right side of system (1)may be represented as
..
{
+
-
where
W=
7
2654
i = 1, ...,n
(3)
Eq.(2) indicates that the dynamics of the i - th neuron in the network are associated with all states
of the network, the synaptic weights w ; , ~ , .. .,Wi,n and the input s;.
2.2 Gersgorin's Theorem Based Stability Conditions
The equilibrium points of the system (1) are defined by the following nonlinear algebraic equation
zf = f ( z f , W , s )
(4)
Without a loss of generality, one can assume that there exist at least one solution of the equation (3);
that is, the system (1) has at least one equilibrium point zf.
Gersgorin's theorem were often used to derive the stability conditions. For a known real or complex matrix, Gersgorin's theorem provides an effective approach for determining the position (of the
eigenvalues of the matrix. In order to analyse the stability of the equilibrium points of system (l), let
d f /dx be the Jacobian of the function f (x,W, s) with respect to z. Using Lyapunov's first method, if
then, zf is a local state equilibrium point. Furthermore, if the elements of the Jacobian df/& are
uniformly bounded, then there exist functions g;,j(wi) gi,j(w;, s;) such that
In this case, if
The equilibrium point xf is a unique global or absolute stable equilibrium point of the system (1).
3. GENERALIZED DBP LEARNING ALGORITHM
In this section, a general formulation of dynamic equilibrium point learning algorithm is developed
for the target pattern storage purpose. Let t = [tl,. ..,t,IT be an output target pattern which is
desired to be implemented by an equilibrium point zf of the neural system (1); that is, t = h(zf).
The purpose of learning procedure is to adjust the synaptic weights w;,, such that ti is be realized by
the nonlinear function h;(zf). Hence, an error function is defined by
n
C[tl- h1(zf)I2= -2l C(Jf)'
1,
2 1=1
E =-
1=1
where Jf = tl - h l ( z f ) . Next, we are going to discuss the learning formulations of the synaptic weights.
After performing a gradient descent in E, the incremental change term of the weight w;,j is given as
where q is a learning rate associated with the synaptic weights. Through a complicated derivation,
the synaptic learning formulations are obtained as
wi,j(k
+ 1)
=
wi,j(k)
af;
+ A w i , j ( k ) = wi,j(k) + qz{(k)=(k)
2655
(10)
where zf and z{ are the equilibrium points of the following dynamic system
Xi(k
+ 1)
= f;(z(k), w;,
Si)
(11)
where J ; ( k ) = t; - x ; ( h ) . Eq.(12) is said to be the adjoint equation associated with the Eq.(ll). The
updating rule (10) is not able to guarantee the stability for both systems (11) and (12), thus, when
Eqs.(ll) and (12) do not converge to the stable equilibrium points xf and z f , the iterative process
needs to be repeated throughly adjusting the learning rate q , until the solutions of Eqs,(ll) and (12)
converge to the stable equilibrium points as time k becomes large.
4. CONSTRAINED LEARNING RATE METHOD
The learning algorithm provided in last section performs a storage procedure in which the target
pattern is stored at an equilibrium point of the neural system (1). In fact, the target pattern is desired
to be stored at an stable equilibrium point of the neural system (1); that is, after finishing the learning
procedure, the synaptic weights would satisfy the stability conditions (5) or (6) so that the stored
pattern is an at least stable equilibrium point. For this goal, one may require that the systems (11)
and (12) are global stable at each learning instant, or after finishing the dynamic learning procedure,
the neural system with the learned synaptic parameters is stable in the sense of Lyapunov.
Let the system at the iterative time k satisfy the row sum stability condition; that is
n
Igi,j(wi(k))l
< 1, i = 1,2,. ..,n
(13)
j=1
We consider the stability of the system at the iterative time k t 1. Let
2
lgi,j(wi(k) t Aw;(k
t 1))l < 1,
i = 1 , 2 , . ..,n
(14)
j=1
Expending the terms on the left side of the above inequality t o second order terms of the incremental
term Aw;(k) produces
It is necessary to note that the iterative time k is omitted in above and following formulations.
Moreover, one may denote that
Then, the learning rate may be requested t o satisfy the second-order algebraic equation for a stable
learning purpose ; that is, a2,;q2t a1,;q t ao,; < 0, i = 1,2,. ..,n. Since ao,; < 0, and Q,;, all; > 0
imply U:,; - 4a2,juo,i > 0, thus, for the positive learning rate q > 0, the solution of the inequality is
obtained as
If one sets
qmin
= min{
-q;+ JU:,; - 4 a o , i ~ , i
2a2,;
Then, the stable learning rate q is obtained as 0 < q
2656
< qmin.
1
5. EXPERIMENTAL RESULTS
In this section, a binary image pattern storage process will be discussed to illustrate the applicability
of the algorithm, where the target pattern is 10 x 10 binary image as shown in Figure la. The equations
of the dynamic network have the following simple form
where the neural activation function a(.)was chosen as a(.)
were easily derived as
n
= tanh(u). The global stability conditions
n
All the iterative initial values of the weights were set t o zeros. Figure 1 shows the equilibrium point
patterns of the network in the iterative process. After 400 steps of the learning, the binary pattern as
shown in Figure l a was stored perfectly at the stable equilibrium point of the network as depicted in
Figure lj.
(f
Figure 1. The binary patterns correspond t o the equilibrium point of the network duriing the
learning process. (a). The target pattern; (b). k=O; (c). k=50; (d). k=100; (e). k=150; (f). k;=200;
(g). k=250; (h). k=300; (i). k=350; (j). k=400.
6. CONCLUSIONS
We have developed a new stable dynamic backpropagation learning algorithm which provides an
associative memory model for real valued vectors using discrete-time dynamic neural model. I n such
a dynamic learning phase, the stability of the network is ensured by a suitable learning rate and,
therefore, the standard DBP routines can be used.
References
[l] F. J, Pineda,“Generalisation of back-propagation to recurrent neural networks,” Pbycical Review Letters, ‘Vol. 59,
No. 19, pp. 2229-2232, 1987.
[2] F.J. Pineda, “Dynamics and architecture for neural computation,” J. of Complexity, Vol. 4, pp. 216-245, 1!)88.
[3] Y. Chauvin, “Dynamic behavior of constrained back-propagation networks,” in Advances in Neural Information
Processing System, D.A. Touretzky, Ed., San Mateo, CA: Morgan-Kaufmann, Vol. 2, pp. 519-526, 1990.
2657