Adaptive multi-channel least mean square and Newton algorithms

Signal Processing 82 (2002) 1127 – 1138
www.elsevier.com/locate/sigpro
Adaptive multi-channel least mean square and Newton
algorithms for blind channel identi&cation
Yiteng Arden Huanga; ∗ , Jacob Benestyb
a Bell
b Bell
Laboratories, Lucent Technologies, Room 2D-526, 600 Mountain Avenue, Murray Hill, NJ 07974, USA
Laboratories, Lucent Technologies, Room 2D-518, 600 Mountain Avenue, Murray Hill, NJ 07974, USA
Received 28 September 2001; accepted 28 March 2002
Abstract
The problem of identifying a single-input multiple-output FIR system without a training signal, the so-called blind system
identi&cation, is addressed and two multi-channel adaptive approaches, least mean square and Newton algorithms, are
proposed. In contrast to the existing batch blind channel identi&cation schemes, the proposed algorithms construct an error
signal based on the cross relations between di4erent channels in a novel, systematic way. The corresponding cost (error)
function is easy to manipulate and facilitates the use of adaptive &ltering methods for an e5cient blind channel identi&cation
scheme. It is theoretically shown and empirically demonstrated by numerical studies that the proposed algorithms converge
in the mean to the desired channel impulse responses for an identi&able system. ? 2002 Published by Elsevier Science B.V.
Keywords: Blind channel identi&cation; Adaptive &ltering; Least mean square; Newton’s method; Multi-channel system
1. Introduction
The desire for blind channel identi&cation and estimation technique arises from a variety of potential
applications in signal processing and communications,
e.g. dereverberation, separation of speech from multiple sources, time-delay estimation, speech enhancement, image deblurring, wireless communications, etc.
In all these applications, a priori knowledge of the
source signal is either inaccessible or very expensive
to acquire, making the blind method a necessity.
Blind channel identi&cation and equalization techniques have gained extensive attention since the innovative idea was &rst proposed by Sato [12]. Early
studies [4,16,2] of blind channel identi&cation and
∗
Corresponding author.
E-mail addresses: [email protected]
Huang), [email protected] (J. Benesty).
(Y.A.
equalization focused primarily on higher (than second) order statistics-based schemes. These schemes
su4er from slow convergence and local minima,
and therefore, are unsatisfactory in tracking a fast
time-varying system. In 1991, Tong et al. [15] demonstrated the possibility of using only second-order
statistics of multi-channel system outputs to solve
the channel identi&cation problem. Since then, many
second order statistics-based approaches have been
proposed, such as the subspace (SS) algorithm [11],
the cross relation (CR) algorithm [8], [17], the least
squares component normalization (LSCN) algorithm
[1], the linear prediction-based subspace (LP-SS) algorithm [13], and the two-step maximum likelihood
(TSML) algorithm [6]. These batch methods can
accurately estimate an identi&able multi-channel system (or equivalently an oversampled single-channel
system [8]) using a &nite number of samples when
the signal-to-noise ratio (SNR) is high. However,
0165-1684/02/$ - see front matter ? 2002 Published by Elsevier Science B.V.
PII: S 0 1 6 5 - 1 6 8 4 ( 0 2 ) 0 0 2 4 7 - 5
1128
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
while these algorithms are able to yield a good estimate of the channel impulse responses, they are in
general computationally intensive and are di5cult to
implement in an adaptive mode [14].
For blind channel identi&cation to be practically
useful in real-time applications, it is imperative that
the algorithm should be computationally simple and
can be adaptively implemented. One e4ort was made
to develop an adaptive algorithm using a neural network [3]. Another attempt was based on the least
squares smoothing (LSS) algorithm [18], which is recursive in order and can be implemented in part using
a lattice &lter. To the best of our knowledge, existing
adaptive blind channel identi&cation algorithms are
algebraically complicated in development and computationally demanding in operation. These characteristics not only obstruct research e4orts for performance
improvement but also make these algorithms unattractive in practical implementations.
In this paper, we approach the problem by formulating a new error function for the outputs of
a multi-channel system based on the CR between
di4erent channels. The proposed error signal then
allows use of many traditional and e5cient adaptive
&lters, in both the time and the frequency domains, in
blind channel identi&cation. For a clear presentation
and an easy performance analysis, we present in this
paper a multi-channel least mean square (MCLMS)
algorithm, which is a generalization of the adaptive eigenvalue decomposition algorithm [5], and a
multi-channel Newton (MCN) algorithm.
2. Problem formulation
The problem addressed in this paper is to determine the impulse responses of a single-input
multiple-output (SIMO) FIR system in a blind way,
i.e. only the observed system outputs are available
and used without assuming knowledge of the speci&c
input signal.
2.1. Channel model
For a multi-channel FIR system as presented in
Fig. 1, the ith observation xi (n) is the result of a linear convolution between the source signal s(n) and
Input
Channels
.
s(n)
Additive Observations
Noise
b 1 (n)
x 1 (n)
+
H1 (z)
b 2 (n)
x 2 (n)
+
H2 (z)
.
.
.
..
.
bM (n)
x M (n)
+
HM (z)
Fig. 1. Illustration of the relationships between the input s(n) and
the observations xi (n) in a single-input multi-channel FIR system.
the corresponding channel response hi , corrupted by
an additive noise bi (n):
xi (n) = hi ∗ s(n) + bi (n);
i = 1; 2; : : : ; M;
(1)
where the ∗ symbol is the linear convolution operator
and M is the number of channels. In a vector form,
the relationship of the input and the observation for
the ith channel is written as
xi (n) = Hi · s(n) + bi (n);
(2)
where
xi (n) = [xi (n) xi (n − 1) · · · xi (n − L + 1)]T ;

hi; 0
 0

Hi = 
 ..
 .
0
hi; 1 · · · hi; L−1 0
hi; 0 · · · hi; L−2 hi; L−1
..
.. ..
..
. .
.
.
· · · 0 hi; 0 hi; 1
···
0
···
..
.
0
..
.



;


· · · hi; L−1
s(n) = [s(n) s(n − 1) · · · s(n − L + 1)
· · · s(n − 2L + 2)]T ;
bi (n) = [bi (n) bi (n − 1) · · · bi (n − L + 1)]T ;
and (·)T denotes a vector=matrix transpose.
The additive noise components in di4erent channels are assumed to be uncorrelated with the source
signal even though they might be mutually dependent. The channel parameter matrix Hi is of dimension L×(2L−1) and is constructed from the channel’s
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
impulse response:
hi = [hi; 0
hi; 1
···
hi; L−1 ]T ;
(3)
where L is set to the length of the longest channel
impulse response by assumption.
A global system equation can be constructed by
concatenating the M single-channel outputs of (2) as
follows:
x(n) = H · s(n) + b(n);
(4)
where
x(n) = [x1T (n)
x2T (n)
···
T
xM
(n)]T ;
T T
H = [H1T H2T · · · HM
] ;
T
T
T
(n) :
b(n) = b1 (n) b2T (n) · · · bM
Therefore, mathematically speaking, the blind
multi-channel identi&cation problem is to estimate
the channel parameter matrix H from the observation
x(n) without using the source signal s(n).
2.2. Channel identi8ability
Before we develop approaches to the blind channel
identi&cation problem, it is worthwhile to discuss the
issues of identi8ability, i.e. if the channels are identi&able or whether the channel impulse responses can be
estimated. A multi-channel FIR system can be blindly
identi&ed mainly because of the channel diversity.
As an extreme counter-example, if all channels are
identical, the system reduces to a single-channel case
and therefore, becomes unidenti&able. In addition,
the source signal needs to have su5cient modes to
make the channels fully excited (as evidence to fully
supported channels). According to [7], two inductive
conditions are necessary and su5cient to ensure system identi&ability, which are shared by all second
order statistics-based blind channel identi&cation
methods:
(1) the polynomials formed from hi ; i = 1; 2; : : : ; M ,
are co-prime, i.e. the channel transfer functions
Hi (z) do not share any common zeros;
(2) the autocorrelation matrix Rss = E{s(n)sT (n)}
of the source signal is of full rank, where E{·}
denotes mathematical expectation.
1129
In the rest of this paper, these two conditions are
assumed to hold so that we are dealing with an identi&able multi-channel FIR system.
3. Adaptive MCLMS and Newton algorithms
3.1. Principle
When the input signal is unknown, the CR between
the sensor outputs can be exploited to estimate the
channel impulse responses. By following the fact:
xi ∗ h j = s ∗ h i ∗ h j = x j ∗ h i ;
i; j = 1; 2; : : : ; M; i = j;
(5)
in the absence of noise, we have the following relation
at time n:
xiT (n)hj = xjT (n)hi ;
i; j = 1; 2; : : : ; M; i = j:
(6)
Multiplying (6) by xi (n) and taking expectation
yields,
R x i x i hj = R x i x j hi ;
i; j = 1; 2; : : : ; M; i = j;
(7)
where Rxi xj = E{xi (n)xjT (n)}. Formula (7) speci&es
M (M − 1) distinct equations. If we sum up the M − 1
CR associated with one particular channel hj , we get
M
Rx i x i hj =
i=1; i=j
M
Rx i x j hi ;
j = 1; 2; : : : ; M:
(8)
i=1; i=j
Over all channels, we then have a total of M equations.
In a matrix form, this set of equations is written as:
Rh = 0;
(9)
where

Rxi xi −Rx2 x1 · · · −RxM x1
 i=1


 −Rx1 x2
Rxi xi · · · −RxM x2


i
=
2
R=

..
..
..
..

.
.
.
.


 −Rx x −Rx x · · ·
Rxi xi
1 M
2 M
i=M
and
h = [h1T
h2T
···
T T
hM
] :













1130
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
In the blind multi-channel identi&cation problem,
Eq. (9) has to be solved for the unknown vector h that
contains all the channel impulse responses. For an
identi&able system in which the two presumptive conditions of Section 2.2 are valid, the matrix R is of rank
ML − 1 and its null space is one-dimensional. Therefore, the channel estimate will be unique if multiplied
by a non-zero constant. In contrast to the CR method
[17] that uses the observed data directly, we construct
the system equation (9) based on the covariance matrices of channel outputs, although the explored subspaces in which the channel impulse response vector
lies are the same.
By examining (9), we see that h is an eigenvector of
R corresponding to a zero eigenvalue. Such a system
equation (9) can be solved for h by direct matrix inversion or eigenvalue decomposition, which lead to a
batch approach. Alternatively, since we are interested
in a more e5cient adaptive algorithm, we will de&ne
an error signal and determine the channel impulse responses by employing a least-squares minimization
strategy.
When observation noise is present, the right-hand
side of (9) is no longer zero and an error vector is
produced,
e = Rh:
(10)
This error can be used to de&ne a cost function
J = e2 = eT e:
(11)
We can then determine a vector ĥ as the solution by
minimizing the cost function (11) in the least-squares
sense:
ĥ = arg min J = arg min hT RT Rh:
h
h
(12)
In this case with the presence of observation noise,
the matrix R is positive de&nite rather than positive
semide8nite and the desired solution ĥ will be the
eigenvector of R corresponding to its smallest eigenvalue.
In order to avoid a trivial estimate with all zero
elements, a constraint has to be imposed on h.
Two constraints have been proposed. One is the
unit-norm constraint, i.e. h = 1. The other is the
component-normalization constraint [1], i.e. cT h = 1,
where c is a constant vector. As an example, if
we know that one element, say hi; k , of the vector
h is equal to which is not zero, then we may
properly specify c = [0; : : : ; 1=; : : : ; 0]T with 1= being the (iL + k)-th element of c. Even though the
component-normalization constraint may be more
robust to noise than the unit-norm constraint [1], a
proper location of the constrained component hi; k
and its value are di5cult to determine a priori in
practice. So a unit-norm constraint will be used in
this paper. Therefore, the blind multi-channel identi&cation is the minimization problem given by (12)
subject to h = 1.
3.2. Multi-channel LMS algorithm
As discussed in the foregoing section and given by
(9), for an identi&able multi-channel FIR system the
unit-norm constraint leads to a solution ĥ that is the
eigenvector of the correlation matrix R corresponding
to a zero eigenvalue in the absence of the observation noise. In practice where noise is always present,
the desired system vector ĥ, which consists of all the
channel impulse responses, is the eigenvector of R
corresponding to the smallest eigenvalue. In order to
estimate the channel impulse responses e5ciently, we
present here an adaptive LMS approach, which is a
generalization of the adaptive eigenvalue decomposition algorithm employed in [7] from a two-channel to
an M -channel (M ¿ 2) system.
3.2.1. Algorithm derivation
To begin, we use (6) to de&ne an error signal based
on the ith and jth observations at time n:
eij (n)
=
xiT (n)hj − xjT (n)hi ; i = j; i; j = 1; 2; : : : ; M;
0;
i = j; i; j = 1; 2; : : : ; M:
(13)
Here, we have (M − 1)M=2 distinct error signals
eij (n), which exclude the case eii (n) = 0 and count the
eij (n) = −eji (n) pair only once. Assuming that these
error signals are equally important, we now de&ne a
cost function as follows:
(n) =
M
−1
M
i=1 j=i+1
eij2 (n):
(14)
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
If the unit-norm constraint h = 1 is enforced at all
times, the error signal becomes:
ij (n)
 T
T

 xi (n)hj =h − xj (n)hi =h;
=
i = j; i; j = 1; 2; : : : ; M;


0;
i = j; i; j = 1; 2; : : : ; M;
=
eij (n)
h(n)
(15)
and the corresponding cost function is given by
J (n) =
M
−1
M
ij2 (n) =
i=1 j=i+1
(k = 1; 2; : : : ; M ) channel impulse response:
M −1 M
2
@(n) @[ i=1
j=i+1 eij (n)]
=
@hk
@hk
=
k−1
=
k−1
2ekj (n)(−xj (n))
j=k+1
M
2eik xi (n) +
i=1
M
M
2eik xi (n) +
i=1
=
2ejk (n)xj (n)
j=k+1
2eik (n)xi (n);
(20)
i=1
(n)
:
h2
(16)
Therefore, the desired solution for h is determined by
minimizing the mean value of the cost function J (n):
ĥ = arg min E{J (n)}:
(17)
h
Direct minimization is computationally intensive and
may be even intractable when the channel impulse responses are long and the number of channels is large.
Here, an LMS algorithm is proposed to solve this minimization problem e5ciently:
ĥ(n + 1) = ĥ(n) − ∇J (n)|h=h(n)
ˆ ;
(18)
where is a small positive step size and ∇ is a gradient
operator.
In order to determine the gradient in (18), we take
a derivative of J (n) with respect to h:
@J (n)
@ (n)
∇J (n) =
=
@h
@h h2
@ (n)
=
@h hT h
1
@(n)
=
− 2J (n)h ;
(19)
h2
@h
where the last step follows from the fact ekk (n)=0. We
may express this equation concisely in matrix form as
follows:
@(n)
= 2X(n)ek (n)
@hk
(21)
= 2X(n)[Ck (n) − Dk (n)]h;
where we have de&ned, for convenience,
X(n) = [x1 (n)
x2 (n)
···
ek (n) = [e1k (n) e2k (n) · · ·
 T
x1 (n)hk − xkT (n)h1
 xT (n)h − xT (n)h
k
2
 2
k
=
..


.
xM (n)]L×M ;
eMk (n)]T



;


T
(n)hk − xkT (n)hM
xM
= [Ck (n) − Dk (n)]h


0 · · · 0 x1T (n) 0 · · · 0
 0 · · · 0 xT (n) 0 · · · 0 


2


 ..
..
..
..
.. 

 . ··· .
.
.
·
·
·
.
Ck (n) = 



T
 0 · · · 0 xM (n) 0 · · · 0 




(k−1)L
(M −k)L
;
M ×ML
T
where
@(n)
=
@h
1131
@(n)
@h1
T @(n)
@h2
T
···
@(n)
@hM
T T
:
Let us now evaluate the partial derivative of (n)
only with respect to the coe5cients of the kth
= [0M ×(k−1)L X (n) 0M ×(M −k)L ]M ×ML ;

 T
xk (n) 0 · · · 0
 0 xT (n) · · · 0 


k

Dk (n) = 
:
..
.. 
 ..
..
.
 .
.
. 
0
0
· · · xkT (n)
M ×ML
1132
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
Continuing, we evaluate the two matrix products in
(21) individually as follows:
Finally, we substitute (26) into (18) and have the update equation
X(n)Ck (n)
ĥ(n + 1)
= [x1 (n)
x2 (n)
···
= ĥ(n) −
xM (n)]L×M
[0M ×(k−1)L XT (n) 0M ×(M −k)L ]M ×ML
M
R̃xi xi (n) 0L×(M −k)L ;
= 0L×(k−1)L
(22)
X(n)Dk (n)
x2 (n)xkT (n)
···
xM (n)xkT (n)]
= [R̃x1 xk (n) R̃x2 xk (n) · · · R̃xM xk (n)];
(23)
where
R̃xi xj (n) = xi (n)xjT (n);
i; j = 1; 2; : : : ; M:
Next, substituting (22) and (23) into (21) yields
@(n)
= 2 −R̃x1 xk (n) − R̃x2 xk (n)
@hk
···

R̃xi xi (n) · · · − R̃xM xk (n) h: (24)
i=k
Thereafter, we incorporate (24) into (19) and obtain
@(n)
= 2R̃(n)h;
@h
∇J (n) =
where
(25)
1
[2R̃(n)h − 2J (n)h];
h2

(28)
If the channel estimate is always normalized after each
update, then we have the simpli&ed algorithm
i=1
= [x1 (n)xkT (n)
2
[R̃(n)ĥ(n) − J (n)ĥ(n)]:
ĥ(n)2
(26)
R̃xi xi (n) −R̃x2 x1 (n) · · · −R̃xM x1 (n)

 i=1





 −R̃ (n)
R̃xi xi (n) · · · −R̃xM x2 (n) 
x1 x2




i=2

:
R̃(n) = 



..
..
.
..
..


.
.
.




 −R̃x x (n) −R̃x x (n) · · ·
R̃xi xi (n) 
1 M
2 M
i=M
(27)
ĥ(n + 1) =
ĥ(n) − 2[R̃(n)ĥ(n) − (n)ĥ(n)]
:
ĥ(n) − 2[R̃(n)ĥ(n) − (n)ĥ(n)]
(29)
The MCLMS adaptive algorithm for blind channel
identi&cation is summarized in Table 1.
3.2.2. Convergence analysis
Assuming that the independence assumption [5]
holds, it can be easily shown that the LMS algorithm
converges in the mean if the step size satis&es
1
0¡¡
;
(30)
max
where max is the largest eigenvalue of the matrix
E{R̃(n) − J (n)IML×ML }.
After convergence, taking the expectation of (28)
gives
R
ĥ(∞)
ĥ(∞)
= E{J (∞)}
;
ĥ(∞)
ĥ(∞)
(31)
which is the desired result: ĥ converges in the mean
to the eigenvector of R corresponding to the smallest
eigenvalue E{J (∞)}.
3.3. MCN algorithm
In the previous section, a MCLMS algorithm was
developed to blindly identify an SIMO FIR system
by minimizing the power of the error signal given by
(13). While the MCLMS algorithm has been shown
to converge in the mean to the desired channel impulse responses, one of the di5culties in the design
and implementation of the LMS adaptive &lter is the
selection of the step size . In selecting the step size
in an LMS algorithm, there is a tradeo4, as pointed
out in many studies, between the rate of convergence,
the amount of excess mean-square error, and the ability of the algorithm to track the system as its impulse
responses change.
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
1133
Table 1
The multi-channel LMS adaptive algorithm for the blind identi&cation of a SIMO FIR system
T
T
T
Parameters:
ĥ = [ĥ1 ĥ2 · · · ĥM ]T , adaptive &lter coe5cients;
, step size
Initialization:
ĥi (0) = [1 0 · · · 0]T ; i = 1; 2; : : : ; M
√
ĥ(0) = ĥ(0)= M (normalization)
Computation:
For n = 0; 1; : : : compute
(a) eij (n) =
(b) (n) =
xiT (n)hj − xjT (n)hi ; i = j; i; j = 1; 2; : : : ; M
0;
i = j; i; j = 1; 2; : : : ; M
M −1 M
i=1
2
j=i+1 eij (n);
(c) Construct the matrix R̃(n) given by (27);
(d) ĥ(n + 1) =
Aiming to achieve a good balance of the three
competing design objectives, we present here a MCN
algorithm (see [9] for the Newton method) with a
variable step size during adaptation:
ĥn+1 = ĥn − E −1 {∇2 J (n)}∇J (n)|h=hˆn ;
(33)
we obtain
∇2 J (n)
=
−
.
Taking mathematical expectation of (35) and invoking
the independence assumption [5] produces
E{∇2 J (n)} = 2R − 4hhT R − 4RhhT
− 2E{J (n)}[IML×ML − 4hhT ]:
(32)
where ∇2 J (n) is the Hessian matrix of J (n) with respect to h. Taking derivative of (26) with respect to h
and using the formula
T
@
@J (n)
+ J (n)IML×ML
[J (n)h] = h
@h
@h
= h[∇J (n)]T + J (n)IML×ML ;
ĥ(n) − 2[R̃(n)ĥ(n) − (n)ĥ(n)]
ĥ(n) − 2[R̃(n)ĥ(n) − (n)ĥ(n)]
(36)
In practice, R and E{J (n)} are not known such that
we have to estimate their values. Since J (n) decreases
as adaptation proceeds and is relative small particularly after convergence, we can neglect the term
E{J (n)} in (36) for simplicity and with appropriate
accuracy, as suggested by simulations. The matrix
R is estimated recursively in a conventional way as
follows:
M
x T xi
i
R̂(0) =
IML×ML ;
L
i=1
2h2 {R̃(n) − [h(∇J (n))T + J (n)IML×ML ]}
h4
4[R̃(n)h − J (n)h]hT
:
h4
(34)
With the unit-norm constraint h = 1, Eq. (34) can
be simpli&ed as follows:
∇2 J (n) = 2{R̃(n) − h[∇J (n)]T − J (n)IML×ML }
T
− 4[R̃(n)h − J (n)h]h :
(35)
R̂(n) = R̂(n − 1) + R̃(n)
for n ¿ 1;
(37)
where (0 ¡ ¡ 1) is an exponential forgetting
factor.
By using these approximations, we &nally get an
estimate W(n) for the mean Hessian matrix of J (n)
and hence deduce the MCN algorithm:
W(n) = R̂(n) − 2ĥ(n)ĥ(n)T R̂(n) − 2R̂(n)ĥ(n)ĥ(n)T ;
(38)
1134
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
Table 2
The multi-channel Newton algorithm for the blind identi&cation of a SIMO FIR system.
T
T
T
T
Parameters:
ĥ = ĥ1 ĥ2 · · · ĥM
, step size
, adaptive &lter coe5cients;
Initialization:
ĥi (0) = [1 0 · · · 0]T ; i = 1; 2; : : : ; M
√
ĥ(0) = ĥ(0)= M (normalization)
Computation:
For n = 0; 1; : : : compute
(a) eij (n) =
(b) (n) =
xiT (n)hj − xjT (n)hi ; i = j; i; j = 1; 2; : : : ; M
0;
i = j; i; j = 1; 2; : : : ; M
M −1 M
j=i+1
i=1
eij2 (n);
(c) Construct the matrix R̃(n) given by (27);
(d) R̂(n) =
 M
x T xi


i

IML×ML ; n = 0





i=1
R̂(n − 1) + R̃(n);
(f) W(n) = R̂(n) −
(g) ĥ(n + 1) =
ĥ(n + 1)
=
ĥ(n) − W−1 (n)[R̃(n)ĥ(n) − (n)ĥ(n)]
;
ĥ(n) − W−1 (n)[R̃(n)ĥ(n) − (n)ĥ(n)]
(39)
where is a new step size, close to but ¡ 1. The MCN
algorithm for blind channel identi&cation is summarized in Table 2.
4. Simulations
We have proposed two adaptive algorithms, the
MCLMS and the MCN algorithms, for the blind channel identi&cation problem and have shown that they
would converge in the mean to the desired channel impulse responses of an identi&able multi-channel system. In this section, we evaluate their performance via
Monte Carlo simulations and demonstrate how they
behave using short channel impulse responses that are
common in digital communication problems. For comparison, the CR algorithm is also studied.
L
2ĥ(n)ĥ(n)T R̂(n)
n¿1
− 2R̂(n)ĥ(n)ĥ(n)T ;
ĥ(n) − W−1 (n)[R̃(n)ĥ(n) − (n)ĥ(n)]
.
ĥ(n) − W−1 (n)[R̃(n)ĥ(n) − (n)ĥ(n)]
The normalized root mean square error (NRMSE)
in dB is used as a performance measure in this paper
and is given by,


N
1
1
U(i) 2  ; (40)
NRMSE , 20 log10 
h N
i=1
where N is the number of Monte Carlo runs, (·)(i)
denotes a value obtained for the ith run, and
hT ĥ
U = h − T ĥ
ĥ ĥ
is a projection error vector. By projecting h onto ĥ and
de&ning a projection error, we take only the misalignment of the channel estimate into account [10].
In all simulations, the source signal was an uncorrelated binary phase shift keying sequence and the
additive noise is i.i.d. zero mean Gaussian. The speci&ed SNR is de&ned as:
2 h2
SNR , 10 log10 s 2 ;
(41)
Mb
where s2 and b2 are the signal and noise powers,
respectively.
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
20
0
+
o
x
NRMSE (dB)
− 20
− 30
− 40
− 50
0
− 10
− 20
− 30
− 40
− 60
− 70
MCLMS
MCN
10
Cost Function J(n) (dB)
− 10
CR
MCLMS
MCN
1135
− 50
0
10
20
30
40
50
SNR (dB)
1]T ;
h2 = [1 − 2 cos( + ) 1]T ;
(42)
where is the absolute phase value of the zeros of the
&rst channel and speci&es the angular distance between the zeros of the two channels on the unit circle.
The use of such a second-order two-channel system
was &rst introduced in [6] and was also employed in
[18,1] This system allows us to investigate the robustness of a blind channel identi&cation algorithm to the
channel identi&ability, conceptually described by the
parameter .
In this simulation, N = 100 Monte Carlo runs were
performed. For the CR method, 50 samples of observations from each channel were used. For the proposed MCLMS and MCN algorithms, the step sizes
= 0:01 and = 0:95 were &xed, respectively. First, a
well-conditioned system is considered with = =10
and = . A comparison of the NRMSEs among the
CR, the MCLMS, and the MCN algorithms is presented in Fig. 2. In this case, the NRMSEs of all
studied algorithms decrease steadily as the SNR increases. The trajectories of the cost function J (n) for
one typical run of the MCLMS and MCN algorithms
at 20 and 50 dB SNRs are presented, respectively, in
100
150
Time n (sample)
200
250
MCLMS
MCN
10
Cost Function J(n) (dB)
h1 = [1 − 2 cos()
50
20
Fig. 2. Comparison of NRMSE vs. SNR among the CR, MCLMS,
and MCN algorithms for the well-conditioned two-channel system
( = =10; = ).
The &rst simulation is concerned with a two-channel
FIR system whose impulse responses in each channel
are second order and are given by:
0
(a)
0
− 10
− 20
− 30
− 40
− 50
(b)
0
50
100
150
Time n (sample)
200
250
Fig. 3. Comparison of convergence between the MCLMS and
MCN adaptive algorithms for the well-conditioned two-channel
system ( = =10; = ). Trajectories of the cost function J (n)
vs. time for one typical run of the MCLMS and MCN algorithms
at 20 dB (a), and 50 dB (b) SNRs.
Fig. 3 (a) and (b) to compare their convergence. As
clearly shown in these &gures, both the MCLMS and
the MCN algorithms can converge quickly to the desired channel impulse responses at these SNRs for
such a well-conditioned system. Second, we studied
an ill-conditioned system with = =10 and = =10.
The resulting NRMSEs of these algorithms are plotted
in Fig. 4. Remarkablely, the adaptive algorithms are
still able to identify the channels. The MCLMS algorithm is stable since no matrix inversion is performed.
However, it cannot be concluded that the proposed
MCLMS and MCN algorithms perform signi&cantly
better than the CR method particularly when the SNR
1136
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
20
0
+
o
x
NRMSE (dB)
− 20
− 30
− 40
− 50
0
− 10
− 20
− 30
− 40
− 60
− 70
MCLMS
MCN
10
Cost Function J(n) (dB)
− 10
CR
MCLMS
MCN
− 50
0
10
20
30
40
50
SNR (dB)
Fig. 4. Comparison of NRMSE vs. SNR among the CR, MCLMS,
and MCN algorithms for the ill-conditioned two-channel system
( = =10; = =10).
0
100
200
300
Time n (sample)
400
500
Fig. 6. Comparison of convergence between the MCLMS and
MCN adaptive algorithms for the ill-conditioned two-channel system ( = =10; = =10) at 50 dB SNR.
20
SNR=0dB
10
Cost Function J(n) (dB)
SNR=10dB
0
SNR=20dB
− 10
SNR=30dB
− 20
SNR=40dB
− 30
SNR=50dB
− 40
− 50
0
0.4
(a)
0.8
1.2
Time n (×104 sample)
1.6
2
20
SNR=0dB
Cost Function J(n) (dB)
10
− 10
SNR=20dB
− 20
SNR=30dB
− 30
SNR=40dB
− 40
SNR=50dB
− 50
(b)
SNR=10dB
0
0
100
200
300
Time n (sample)
400
500
Fig. 5. Trajectories of the cost function J (n) vs. time for one typical
run of the MCLMS (a) and MCN (b) algorithms at di4erent SNRs
for the ill-conditioned two-channel system ( = =10; = =10).
is ¿35 dB because the adaptive algorithms uses more
samples in this simulation. The trajectories of the cost
function J (n) for one typical run of the NRMSE and
MCN algorithms at di4erent SNRs are presented in
Fig. 5(a) and (b), respectively. A close snapshot of the
trajectories of these adaptive algorithms at 50 dB SNR
for comparison of their convergence is presented in
Fig. 6. We see that the closer the zeros of the two channels approach to each other, the slower the MCLMS
converges. In terms of convergence, the MCN algorithm is more robust to the system identi&ability and
it takes only about 100 iterations to converge at 50 dB
SNR. The convergence is dramatically accelerated by
the MCN algorithm.
In the second experiment, we consider a threechannel system whose impulse responses are longer
(L = 15) than those in the &rst simulation. An FIR
system with more channels is less likely for all of the
channels to share a common zero which would invalidate the system’s identi&ability. This system was &rst
used for performance evaluation in [1]. The coe5cients of the three impulse responses were randomly
selected whose zeros are shown in Fig. 7. The NRMSEs were obtained by averaging the results of N =200
Monte Carlo runs. The CR method used 120 samples
and the step sizes for the MCLMS and MCN algorithms were = 0:001 and = 0:95, respectively. The
results are shown in Figs. 8–10, using the same layout
as that for the ill-conditioned two-channel system. In
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
2.5
30
Channel 1
Channel 2
Channel 3
2.0
1137
SNR=0dB
20
Cost Function J(n) (dB)
Imaginary Part
1.5
1.0
0.5
0
− 0.5
− 1.0
− 1.5
0
SNR=20dB
− 10
SNR=30dB
− 20
SNR=40dB
− 30
SNR=50dB
−3
−2
−1
0
1
2
3
+
o
x
CR
MCLMS
MCN
Cost Function J(n) (dB)
0
− 20
1.5
20
SNR=0dB
10
SNR=10dB
0
SNR=20dB
− 10
SNR=30dB
− 20
SNR=40dB
− 30
− 40
− 30
SNR=50dB
0
(b)
− 40
− 50
0.5
1.0
Time n (×104 sample)
30
Fig. 7. Illustration of the positions of the channel zeros for the
three-channel system in a z-plane.
− 10
0
(a)
Real Part
NRMSE( dB)
SNR=10dB
− 40
− 2.0
− 2.5
10
200
400
600
Time n (sample)
800
1000
Fig. 9. Trajectories of the cost function J (n) vs. time for one
typical run of the MCLMS (a) and MCN (b) algorithms at di4erent
SNRs for the three-channel system.
0
10
20
30
40
50
SNR( dB)
Fig. 8. Comparison of NRMSE vs. SNR among the CR, MCLMS,
and MCN algorithms for the three-channel system.
this case, even though some zeros of the three channels are very close and a small step size was used, the
MCLMS converged faster than for the ill-conditioned
two-channel system. The MCN method performed
still better than the MCLMS algorithm.
As clearly demonstrated by these simulations, the
MCLMS and MCN algorithms are robust to the system
identi&ability and noise. They converge in the mean
to the desired channel impulse responses. Poor convergence rate of the proposed MCLMS algorithm for
ill-conditioned systems is mainly because only an instantaneous gradient estimate is used for each update
in the LMS algorithm. Such a critical drawback has
been overcome by the MCN algorithm which controls
the step size based on the Hessian of the cost function.
In this paper, we formulate the CR between channels
in a systematic way and the &nal concise cost function potentially facilitates the use of adaptive &ltering algorithms, such as the block LMS, the recursive
least squares (RLS), and the frequency-domain LMS
methods, in the future development of e5cient blind
channel identi&cation schemes.
5. Conclusions
The blind channel identi&cation and estimation
problem is studied in this paper. An error function
1138
Y.A. Huang, J. Benesty / Signal Processing 82 (2002) 1127 – 1138
30
MCLMS
MCN
Cost Function J(n) (dB)
20
10
0
− 10
− 20
− 30
− 40
0
200
400
600
Time n (sample)
800
1000
Fig. 10. Comparison of convergence between the MCLMS and
MCN adaptive algorithms for the three-channel system at 50 dB
SNR.
based on the cross relations between di4erent channels
is constructed in a systematic way for a multi-channel
FIR system. The resulting cost function is concise
and potentially facilitates the use of adaptive &ltering techniques in the future development of e5cient
adaptive blind channel identi&cation schemes. As an
example, an LMS approach was proposed and its convergence in the mean to the desired channel impulse
responses was theoretically shown. Furthermore, in
order to accelerate convergence of the LMS adaptive
algorithm, a Newton method was proposed. Simulation results justi&ed our analysis and the proposed
MCLMS and MCN algorithms performed well for an
identi&able system.
Acknowledgements
We would like to thank B.H. Juang and D.R. Morgan for their constructive comments and suggestions
that have improved the clarity of this paper.
References
[1] C. Avendano, J. Benesty, D.R. Morgan, A least squares
component normalization approach to blind channel
identi&cation, in: Proceedings of the IEEE International
Conference Acoustics, Speech, Signal Processing, Vol. 4,
1999, pp. 1797–1800.
[2] A. Benveniste, M. Goursat, Blind equalizers, IEEE Trans.
Commun. COM-32 (8) (August 1984) 871–883.
[3] G. Dong, R.W. Liu, An orthogonal learning rule for
null-space tracking with implementation to blind two-channel
identi&cation, IEEE Trans. Circuits Systems I 45 (1) (January
1998) 26–33.
[4] D.N. Godard, Self-recovering equalization and carrier
tracking in two-dimensional data communication systems,
IEEE Trans. Commun. COM-28 (11) (November 1980)
1867–1875.
[5] S. Haykin, Adaptive Filter Theory, Prentice-Hall, Englewood
Cli4s, NJ, 1996.
[6] Y. Hua, Fast maximum likelihood for blind identi&cation of
multiple FIR channels, IEEE Trans. Signal Process. 44 (3)
(March 1996) 661–672.
[7] Y. Huang, J. Benesty, G.W. Elko, Adaptive eigenvalue
decomposition algorithm for real-time acoustic source
localization system, in: Proceedings of the IEEE International
Conference Acoustics, Speech, Signal Processing, Vol. 2,
1999, pp. 937–940.
[8] H. Liu, G. Xu, L. Tong, A deterministic approach to blind
equalization, in: Proceedings of the 27th Asilomar Conference
on Signals, Systems, and Computers, Vol. 1, 1993, pp.
751–755.
[9] T.K. Moon, W.C. Stirling, Mathematical Methods and
Algorithms, Prentice-Hall, Upper Saddle River, NJ, 1999.
[10] D.R. Morgan, J. Benesty, M.M. Sondhi, On the evaluation
of estimated impulse responses, IEEE Signal Process. Lett.
5 (7) (July 1998) 174–176.
[11] E. Moulines, P. Duhamel, J.F. Cardoso, S. Mayrargue,
Subspace methods for the blind identi&cation of multichannel
FIR &lters, IEEE Trans. Signal Process. 43 (2) (February
1995) 516–525.
[12] Y. Sato, A method of self-recovering equalization for
multilevel amplitude-modulation, IEEE Trans. Commun.
COM-23 (6) (June 1975) 679–682.
[13] D. Slock, Blind fractionally-spaced equalization, perfect
reconstruction &lerbanks, and multilinear prediction, in:
Proceedings of the IEEE International Conference Acoustics,
Speech, Signal Processing, Vol. 4, 1994, pp. 585 –588.
[14] L. Tong, S. Perreau, Multichannel blind identi&cation: from
subspace to maximum likelihood methods, Proc. IEEE 86
(10) (October 1998) 1951–1968.
[15] L. Tong, G. Xu, T. Kailath, A new approach to
blind identi&cation and equalization of multipath channels,
Proceedings of the 25th Asilomar Conference on Signals,
Systems, and Computers, Vol. 2, 1991, pp. 856 –860.
[16] J.R. Treichler, B.G. Agee, A new approach to multipath
correction of constant modulus signals, IEEE Trans. Acoust.
Speech Signal Process. ASSP-31 (2) (April 1983) 459–472.
[17] G. Xu, H. Liu, L. Tong, T. Kailath, A least-squares approach
to blind channel identi&cation, IEEE Trans. Signal Process.
43 (December 1995) 2982–2993.
[18] Q. Zhao, L. Tong, Adaptive blind channel estimation by
least squares smoothing, IEEE Trans. Signal Process. 47 (11)
(November 1999) 3000–3012.