me577ParticleFiltering_Julier.pdf

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
477
Technical Notes and Correspondence_______________________________
A New Method for the Nonlinear Transformation of Means
and Covariances in Filters and Estimators
Simon Julier, Jeffrey Uhlmann, and Hugh F. Durrant-Whyte
Abstract—This paper describes a new approach for generalizing the
Kalman filter to nonlinear systems. A set of samples are used to parameterize the mean and covariance of a (not necessarily Gaussian) probability distribution. The method yields a filter that is more accurate than
an extended Kalman filter (EKF) and easier to implement than an EKF or
a Gauss second-order filter. Its effectiveness is demonstrated using an example.
z(k) = h[x(k); u(k); k] + w(k)
(1)
where x(k) is the state of the system at timestep k; u(k +1) is the input
vector, v(k) is the noise process caused by disturbances and modelling
errors, z(k) is the observation vector, and w(k) is additive measurement noise. It is assumed that the noise vectors v(k) and w(k) are zero
mean and
T
E [v(i)v (j )] = ij Q(i)
T
E [w(i)w (j )] = ij R(i);
T
E [v(i)w (j )] = 0:
8 i; j
Index Terms—Covariance matrices, estimation, filtering, missile detection and tracking, mobile robots, nonlinear filters, prediction methods.
^ (ijj )
The MMSE estimate of x(k) is the conditional mean. Let x
be the mean of x(i) conditioned on all of the observations up to and
including time j
I. INTRODUCTION
x^ (ijj ) = E [x(i)jZj ]
The problem of generalizing the Kalman filter paradigm for nonlinear applications is considered. This work is motivated by the wellknown limitations of the extended Kalman filter (EKF), which simply
linearizes all nonlinear models so that the traditional linear Kalman
filter can be applied. Although the EKF (in its many forms) is a widely
used filtering strategy [1], [2], over 30 years of experience with it has
led to a general consensus within the tracking and control community
that it is difficult to implement, difficult to tune, and only reliable for
systems that are almost linear on the time scale of the update intervals.
As is well known, the optimal solution to the nonlinear filtering
problem is infinite dimensional [3] and a large number of suboptimal
approaches have been developed [4], [5]. These methods can be broadly
classed as numerical Monte Carlo [6] methods or analytical approximations [7]–[9]. However, the application of these methods to high-dimensioned systems is rarely practical, and it is a testament to the conceptual simplicity of the EKF that it is still widely used.
In this paper, a new linear estimator [10]–[12] is developed. It yields
performance equivalent to the Kalman filter for linear systems, yet generalizes elegantly to nonlinear systems without the linearization steps
required by the EKF. We show analytically that the expected performance of the new approach is superior to that of the EKF and, in fact,
lies between those of the modified, truncated second-order filter [13]
and the Gaussian second-order filter [14]. However, the algorithm is not
restricted to Gaussian distributions. We demonstrate the performance
benefits in an example application, and we argue that the ease of implementation and more accurate estimation features of the new filter
recommend its use over the EKF in virtually all applications.
II. BACKGROUND
We seek the minimum-mean-squared error (MMSE) estimate of the
state vector of the nonlinear discrete time system
x(k + 1) = f [x(k); u(k); v(k); k]
where Zj = [z(1); 1 1 1 ; z(j )]: The covariance of this estimate is denoted P(ijj ):
The Kalman filter propagates the first two moments of the distribution of x(k) recursively and has a distinctive “predictor-corrector”
^ (k jk ); the filter first predicts what the
structure. Given an estimate x
future state of the system will be using the process model. Ideally, the
predicted quantities are
x^ (k + 1jk) = E [f [x(k); u(k); v(k); k]jZk ]
P(k + 1jk) = E [fx(k + 1) 0 x^ (k + 1jk)g
2 fx(k + 1) 0 x^(k + 1jk)gT jZk ]:
The expectations can be calculated only if the distribution of x(k)
conditioned on Zk is known. In general, the distribution cannot be described by a finite number of parameters and most practical systems
employ an approximation of some kind. It is conventionally assumed
that the distribution of x(k) is Gaussian at any time k: Two justifications are made. First, only the mean and covariance need to be maintained. Second, given just the first two moments the Gaussian distribution is the entropy maximizing or least informative distribution [15].
The estimate at time k + 1 is given through updating the prediction
by the linear update rule
x^ (k + 1jk + 1) = x^ (k + 1jk) + W(k + 1) (k + 1)
P(k + 1jk + 1) = P(k + 1jk) 0 W(k + 1)P (k + 1jk)
1 WT (k + 1)
^(k + 1jk )
(k + 1) = z(k + 1) 0 z
1
W(k + 1) = Px (k + 1jk)P0
(k + 1jk ):
~ (ijj ) =
The EKF exploits the fact that the error in the prediction, x
x(i) 0 x^ (ijj ); can be attained by expanding (1) as a Taylor Series about
^ (k jk ): Truncating this series at the first order yields the
the estimate x
approximate linear expression for the propagation of state errors as
Manuscript received November 16, 1994; revised December 1, 1995,
September 1, 1996, and June 1, 1998. Recommended by Associate Editor, V.
Solo.
S. Julier is with IDAK Industries, Jefferson City, MO USA (e-mail:
[email protected]).
J. Uhlmann is with the Robotics Research Group, Oxford University, U.K.
H. F. Durrant-Whyte is with the Department of Mechanical and Mechatronic
Engineering, Sydney, Australia (e-mail: [email protected]).
Publisher Item Identifier S 0018-9286(00)02145-0.
x~ (k + 1jk) rfx x~ (kjk) + rfv v(k)
where rfx is the Jacobian of f [1] with respect to x(k); and rfv is that
with respect to v(k): Using this approximation, the state prediction
equations are
x^ (k + 1jk) = f [^x(kjk); u(k); 0]
0018–9286/00$10.00 © 2000 IEEE
478
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
P(k + 1jk) = rfx P(kjk)rT fx + rfv Q(k + 1)rT fv :
n
(2)
This approximation is valid if the contributions of the truncated
higher order terms are negligible. However, in many practical situations (such as converting from polar to Cartesian coordinate systems
[17]), linearization introduces significant biases or errors. The EKF
can also be prohibitively difficult to implement because deriving
Jacobians can be cumbersome and time consuming (especially when
the system is complicated and of high order), and they must be
reevaluated at every prediction step of the filter.
This paper focuses on the central problem of predicting the mean and
covariance. We present an alternative algorithm that is more accurate,
convenient, and uses approximately the same number of calculations
as an EKF.
=
=
P(kjk):
x
P
x
X0 (kjk) = x^(kjk)
W0 = =(n + )
Xi (kjk) = x^(kjk) +
P jk)
(n + ) (k
Wi = 1=f2(n + )g
Xi+n (kjk) = x^(kjk) 0 (n + )P(kjk)
Wi+n = 1=f2(n + )g:
2 <; (
i
i
(3)
P j
P j
(n + ) (k k ))i is the ith row or column1 of the matrix
of (n + ) (k k), and Wi is the weight that is associated
square root
with the ith point.
Theorem 1: The set of samples chosen by (3) have the same sample
mean, covariance, and all higher odd-ordered central moments as the
distribution of (k): The matrix square root and κ affect the fourth and
higher order sample moments of the sigma points.
Proof: The matching of the mean, covariance, and all odd-ordered moments can be readily demonstrated. Because the points are
symmetrically distributed and chosen with equal weights about ; the
sample mean is obviously and all odd-ordered moments are zero. The
sample covariance is
x
P=
=
2n
i=0
n
i=1
P
x
x
2Wi (n + )
A P
P(kjk)
i
P(kjk)
P=A A
T
i
1If the matrix square root
of is of the form
, then the sigma
points are formed from the rows of : However, for a root of the form
, the columns of
are used.
AA
A
A
P=
T
i
2) The predicted mean is computed as
x^(k + 1jk) =
2n
i=0
Wi Xi (k + 1jk):
(4)
3) The predicted covariance is computed as
P(k + 1jk) =
2n
i=0
Wi fXi (k + 1jk) 0 x^ (k + 1jk)g
2 fXi (k + 1jk) 0 x^(k + 1jk)gT :
(5)
The mean and covariance are calculated using standard vector and
matrix operations, which means that the algorithm is suitable for any
choice of process model, and implementation is extremely convenient
because it is not necessary to evaluate the Jacobians, which are needed
in an EKF. The method has a further advantage: it yields more accurate
predictions than those determined through linearization.
Theorem 2: The prediction algorithm introduces errors in estimating the mean and covariance at the fourth and higher orders in the
Taylor series. These higher order terms are a function of κ and the
matrix square root used.
Proof: Appendix II shows that evaluating the mean and covariance correctly up to the mth order requires approximating the moments
of (k) up to the mth order. From Theorem 1, the approximation is correct up to the third order, and so errors are introduced in the fourth and
higher orders. These are a function of the matrix square root and κ.
Remark 3: This is the same order of accuracy as the second-order
Gaussian filter [14], but without the need to calculate Jacobians or Hessians.
provides an extra degree of freedom to “fine tune” the higher order
moments of the approximation, and can be used to reduce the overall
prediction errors. When (k) is assumed Gaussian, Appendix I shows
that a useful heuristic is to select n + = 3: If a different distribution
is assumed for (k), then a different choice of might be more appropriate.
Although can be positive or negative, Appendix III shows that
when is negative it is possible to calculate a nonpositive, semidefinite
(k + 1jk ): This problem is not uncommon for methods that approximate higher order moments or probability density distributions (such
as those described in [5], [4], and [8]). In this situation, it is possible to
use a modified form of the prediction algorithm. The mean is still calculated using (4), but the “covariance” is evaluated about X0 (k + 1jk):
Appendix III shows that the modified form ensures positive semi-definiteness, and, in the limit (n + ) ! 0; the prediction is the same as
that of the modified, truncated second-order filter [4].
x
x
x
P
Wi fXi (kjk) 0 x^ (kjk)gfXi (kjk) 0 x^ (kjk)gT
P(kjk)
Xi (k + 1jk) = f [Xi (kjk); u(k); k]:
x
P
i=1
i
Appendix I sketches the principle of the argument for the higher
moments, and a full proof can be found in [21].
Remark 1: The above properties hold for any choice of the matrix
square root. Efficient and stable methods, such as the Cholesky decomposition, should be used.
Remark 2: κ can be any number (positive or negative) providing
that (n + ) 6= 0:
Given the set of samples generated by (3), the prediction procedure
is as follows.
1) Each sigma point is instantiated through the process model to
yield a set of transformed samples
III. THE NEW FILTER
We use the intuition that it is easier to approximate a probability
distribution than it is to approximate an arbitrary nonlinear function
or transformation [18]. Following this intuition, we generate a set of
points whose sample mean and sample covariance are ^ (kjk) and
(k jk ), respectively. The nonlinear function is applied to each of
these points in turn to yield a transformed sample, and the predicted
mean and covariance are calculated from the transformed sample.
Although this superficially resembles a Monte Carlo method, the
samples are not drawn at random. Rather, the samples are deterministically chosen so that they capture specific information about the distribution. In general, this intuition can be applied to capture many kinds
of information about many types of distribution [19], [20]. In this paper,
we consider the special case of i) capturing the mean and covariance of
an ii) assumed Gaussian distribution.
The n-dimensional random variable (k) with mean ^ (kjk) and covariance (kjk) is approximated by 2n +1 weighted samples or sigma
points selected by the algorithm
P(kjk)
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
479
Remark 4: The method can be generalized to include the effects of
process and observation noise by appending the noise vectors to the
state vector [11].
IV. EXAMPLE APPLICATION
This section applies and compares the performance of the new filter
against an EKF for a tracking problem that was presented in [14]. This
example was chosen because it has significant nonlinearities in the
process and observation models and has been analyzed extensively in
the literature.
We wish to estimate the position x1 (t); velocity x2 (t), and constant
ballistic coefficient x3 (t) of a body as it reenters the atmosphere at
a very high altitude at a very high velocity. Its motion is determined
by altitude- and velocity-dependent drag terms, and it is constrained to
fall vertically. The position of the body is measured at discrete points in
time using a radar capable of measuring range corrupted by Gaussian
measurement noise. The radar is at an altitude of H (100 000 ft), and
the horizontal range between the body and the radar, M; is (100 000
ft).
The continuous time dynamics of this system are
x_ 1 (t) =0x2 (t) + w1 (t)
x_ 2 (t) =0e0x
(t)
Fig. 1.
Absolute mean error position error.
(6)
2
x2 (t) x3 (t) + w2 (t)
x_ 3 (t) =w3 (t)
(7)
(8)
where w1 (t); w2 (t), and w3 (t) are zero-mean, uncorrelated noises
with covariances given by Q(t) and is a constant (5 2 1005 ) that
relates the air density with altitude. The range at time t; z (t); is
z (t) =
(M 2 + [x1 (t)
0 H ] ) + r(t);
2
(9)
where r(t) is the uncorrelated observation noise with covariance
4
2
= 10 ft . The measurements are made with a frequency of 1
Hz.
Tracking systems were implemented using the new filter and the
EKF. The nonlinearities of the process model and the high velocities
required the numerical integration of (6)–(8) to be carried out using
extremely small time steps. In accordance with [14], a fourth-order
Runge–Kutta scheme was employed with 64 steps between each observation. For the EKF, it was necessary to recalculate the Jacobian 64
times between each update. For the new filter, the trajectory of each
sigma point was calculated using the small time steps, but it was only
necessary to calculate the mean and covariance just before an observation was made. Because n = 3; was chosen to be zero in accordance
with the heuristic n + = 3:
The initial true state of the system is x(0) = [3 2 105 2 2 104 1003 ],
and the initial estimates and covariances of these states are
R(t)
x^ (0j0) = [3 2 105 2 2 104 3 2 1005 ]
6
P(0j0) =
10
0
0
0
4
0
2 10
0
0
10
6
:
Although the initial estimates of altitude and velocity are correct,
x^ 3 (0j0) is very bad. The body is assumed to be “heavy,” whereas in
reality it is “light.” The behavior of the two filters differs if the second
and higher order terms are significant. Because process noise can be
used to mask linearization errors, we adopt the practice from [14] and
do not introduce any process noise into the simulation—Q(k) = 0
for both filters.
In Fig. 1, we show the average magnitude of the state errors committed by each filter across a Monte Carlo simulation consisting of 50
Fig. 2. Absolute mean velocity error.
runs. At high altitude, the drag effects are minimal, and the body falls
approximately linearly. However, after about 10 s, drag becomes significant, motion becomes noticeably nonlinear, and the two filters differ
significantly. The velocity estimates are shown in Fig. 2 and indicate
large error spikes in both filters. These occur when the altitude of the
body is the same as that of the radar and range information provides
less data about body motion.The new filter recovers quickly, but the
EKF has a larger error spike and only slowly converges. Fig. 3 shows
the errors in estimating x3 (1): The error in the EKF estimate is biased
and is an order of magnitude larger than that for the new filter.
In Figs. 4 and 5, we show the errors in the position estimates made
by the EKF and the new filter and the associated estimates of the two
standard deviation bounds. These bounds are given by twice the square
root of the diagonals of the covariance matrix, and, if the filter is consistent, the state errors should lie within these bounds 95% of the time.
However, the error in the EKF drifts outside of these bounds after 30 s,
showing that it does not yield consistent estimates. However, the errors
in the new filter always lie well within the two standard deviations, implying that the new filter is consistent.
480
Fig. 3.
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
Absolute mean error in x
:
Fig. 5.
New filter mean position error.
This algorithm has found a number of applications in high-order,
nonlinear coupled systems, including navigation systems for
high-speed road [21], [16] vehicles, public transportation systems
[22], and underwater vehicles [23]. Square root filters can be formulated by discrete sets of points (as demonstrated in [24]), and
iterated filters can be constructed using the predictions [25]. The
algorithm has been extended to capture the first four moments of a
Gaussian distribution [12] and the first three moments of an arbitrary
distribution [20].
Given its performance and implementation advantages, we conclude
that the new filter should be preferred over the EKF in virtually all
nonlinear estimation applications.
APPENDIX I
HIGHER MOMENTS OF THE APPROXIMATION
Fig. 4.
EKF mean position error.
Therefore, we conclude that in this example the new filter has substantial advantages over the EKF both in implementation and performance.
V. CONCLUSIONS
Motivated by the deficiencies of the EKF, we have examined a completely new approach for applying linear estimation theory to nonlinear systems. Rather than approximate the Taylor series to an arbitrary order, we approximate the first three moments of the prior distribution accurately using a set of samples. The algorithm predicts the
mean and covariance accurately up to the third order and, because the
higher order terms in the series are not truncated, it is possible to reduce the errors in the higher order terms as well. We have provided
empirical evidence that supports the theoretical conclusion and have
demonstrated that the new filter is far easier to implement because it
does not involve any linearization steps, eliminating the derivation and
evaluation of Jacobian matrices.
This appendix analyzes the properties of the higher moments of the
sigma point selection scheme of (3). The selection process consists of
three stages. First, a set of sigma points are drawn to approximate an
n-dimensional standard Gaussian with mean 0 and covariance I: A
linear transformation [a matrix square root P(kjk )] is applied to each
point so that the transformed sample still has mean 0, but the covari^ (k jk ) is added to each transformed sigma
ance is P(kjk): Finally, x
point to ensure the correct sample mean. affects the first stage of the
approximation, and the matrix square root affects the second.
The sigma points, which are assumed approximate the standard
Gaussian, lie on the coordinate axes.2 The orthogonality and the symmetry of these points means that the only nonzero sample moments
are those that are an even order power of a single coefficient. The
(k01)
2k th-order moment for any coefficient is (n + )
: Therefore,
the higher order moments scale geometrically by a factor determined
by . However, the moments for the standard Gaussian are different
from those of the sigma points, which approximate it. Because the
covariance matrix is I, the different components are independent of
one another and
E [x1
2In
I:
2 111 2
xn
n
]=
i=1
E [xi
]
effect, these points are drawn from the orthogonal matrix square root of
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
481
where the ith component in the moment is raised to the power i : It
can be shown that [13]
E [xi
] =
i is odd:
;
0
1
2 2 1 1 1 2 i 0 ; i is even:
3
1)
(
Two differences exist in the approximation by the sigma points.
First, the “scaling” of nonzero moments are different. Second, a
number of moment terms are “missed out.” These differences become
more marked as the order increases, but, as explained in Appendix II,
they only affect the higher order terms of the Taylor Series of the
process model/nonlinear function. Assuming that the contribution of
terms diminishes as the order increases, the emphasis is to minimize
the errors in the lowest order terms. Because E [x4i ] = 3 for the
Gaussian distribution, choosing (n + ) = 3 minimizes the difference
between the moments of the standard Gaussian and the sigma points
up to the fourth order.
Because the fourth and higher order moments are not precisely
matched, the choice of matrix square root affects the errors in the
higher order terms by adjusting the way in which the errors are
distributed among the different states. However, in general, this information cannot be exploited because it would require knowledge of the
higher order derivatives of the process model equation. In the absence
of this information, the choice of matrix square root is governed by
other issues, such as numerical stability or computational cost.
=
y
y fx
x x x
P
x
y
x
x
x
x
x
y fx x
2
3
4
f x D1xf D1xf D1xf D1xf 1 1 1 (10)
where the D1x f operator evaluates the total differential of f 1 when
perturbed around a nominal value x by x: The ith term in the Taylor
series for f 1 is given by
i
n
D1i xf
(11)
xj @ f x
=
[
=
[
+ ~]
] +
+
+
2!
+
3!
4!
[ ]
[ ]
=
1
i! j =1
~
@xj
[
y
y
=
=
x
E [f [x + x~ ]]
fx
[
] +
E
2
3
4
D1xf D1xf D1xf D1xf 1 1 1
+
2!
+
+
3!
=
D1i xf
i!
=
1
i!
E
n
j =1
xj @x@ j
~
i
fx
[
]
+ ~]
[
=
111
(13)
+ ~ ]]
[ [
+
+
2!
2!
+
+
3!
4!
(14)
+
4!
with substitutions from (10) and (12). The true covariance is found
by postmultiplying the state error by the transpose of itself and taking
expectations. Exploiting the fact that ~ is symmetric, the odd terms all
evaluate to zero and the covariance is given by
x
Pyy rfxPxxrfxT
3 T D12 x f D12 x f T
E D1x f D1x f
(12)
(
+
+
)
3!
(
+
2
2
D13 xf D1xf T 0 E D12 xf
(
)
2!
E
)
3!
2
D12 xf
T
2
+
111
(15)
D fD f
fP
f
where E [ 1x ( 1x )T ] = r x xx r xT has been used. Comparing
this with (2), it can be seen that the EKF truncates this series after the
first term. The new filter does not truncate the series at any arbitrary
order, and, applying the analysis from above, it is correct up to the
second order with errors in the fourth and higher order terms. Therefore, both the EKF and the new filter predict the covariance correctly
up to the third order.
APPENDIX III
THE MODIFIED FORM OF THE ALGORITHM
When is negative, it is possible that the predicted covariance will
not be positive semidefinite. This can be demonstrated by taking the
limit of (15) as (n + ) ! 0: As shown in Appendix I, the fourth and
higher order moments tend to zero and
D12 xf
T
(n+)!0 Pyy rfx Pxx rfx 0 E
lim
=
2
E
D12 xf
2
T
:
It can be seen that a fourth-order, positive-semidefinite matrix is subtracted. However, this term originates from the outer product of two
second-order expected terms and does not scale with .
The modified form of the algorithm evaluates the covariance about
the projected mean. Taking the sigma points for are Xi and those for
are Yi ; the modified form for calculating yy is3
y
PMOD
yy
where the ith term in the series is given by
E
4!
+
+
y y y y
y0y f x x 0E f x x
2
3
4
D1xf D1xf D1xf D1xf
2
4
0 E D1x f D1x f 1 1 1
]
where x
~j is the j th component of ~ : Therefore, the ith term in the series
is an ith-order polynomial in the coefficients of ~ ; whose coefficients
are given by derivatives of [1]:
is the expected value of (10)
f
2
P
x=X
x
1
x
+
~
i!
f
@i
i
@x 01 @x
m111112
=
In this appendix we provide a justification for approximating a probability distribution by a set of samples that match its moments. We show
that an approximation is only correct to the mth order if its moments
up to that order are correctly approximated.
Consider a Gaussian-distributred random variable with mean
and covariance xx : We wish to calculate the mean and covariance
yy of the random variable , which is related to through the nonlinear (analytic) function = [ ]:
Noting that can be written as = + ~ , where ~ is a zero-mean
Gaussian random variable with covariance xx ; the nonlinear transformation can be expanded as a Taylor Series about
P
i!
+
and mc c 111c is the ith-order moment mc c 111c = E [x
~c x
~c 1 1 1 x
~c ]:
The ith term is a function of the ith central moments of the distribution of ~ : Therefore, for an approximation to be accurate up to the ith
order, it must be able to approximate the moments correctly up to the
ith order. The EKF truncates this series after the first term, and so its
error in predicting the mean is in the second and higher orders. The new
filter matches the mean and covariance correctly, and so it is correct up
to the second order. Because the new filter does not truncate the filter
at any order, it can be expected that the errors in the fourth and higher
order terms are smaller than those committed by the EKF.
The covariance is given by yy = E [f 0 gf 0 gT ]: Now
APPENDIX II
MOMENT APPROXIMATION AND PERFORMANCE
P
i
m111111 @ fi
@x1
1
2n
=
i=1
P
x
Wi [Yi 0 Y0 ][Yi 0 Y0 ]T :
3This is equivalent to calculating the mean and covariance using (4) and (5)
and adding a term fy
y
:
= Y gf 0 Y g
482
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 45, NO. 3, MARCH 2000
Positive semidefiniteness is guaranteed by the fact that the covariance matrix is evaluated as the sum of outer products of vectors. It can
be shown that the calculated covariance is
Pyy
=r
fxPxx fxT
D1xf D31xf T D12 xf D12 xf T
r
(
+ E
)
+
D13 xf D1xf T
(
(
+
3!
)
)
2 2 2!
+ 111
3!
which introduces errors in the fourth and higher orders.
The modified form has the advantage that, irrespective of the choice
of the matrix square root
D12 xf
y
f
x
(n+)!0
fxPxx fxT
(n+)!0 Pyy
lim
lim
=
[
] + E
=r
r
2!
:
These are the values calculated by the modified, truncated secondorder filter [4], but without the need to evaluate Jacobians or Hessians.4
ACKNOWLEDGMENT
The authors would like to thank R. Bellaire at Georgia Tech for insightful comments on an early draft of this paper and the referees for
their many constructive remarks.
REFERENCES
[1] J. K. Uhlmann, “Algorithms for multiple target tracking,” Am. Sci., vol.
80, no. 2, pp. 128–141, 1992.
[2] H. W. Sorenson, Kalman Filtering: Theory and Application. New
York: IEEE Press, 1985.
[3] H. J. Kushner, “Dynamical equations for optimum non-linear filtering,”
J. Differential Equations, vol. 3, pp. 179–190, 1967.
[4] P. S. Maybeck, Stochastic Models, Estimation, and Control, P. S. Maybeck, Ed. New York: Academic, 1982, vol. 2.
[5] A. H. Jazwinski, Stochastic Processes and Filtering Theory. New
York: Academic, 1970.
[6] N. J. Gordon, D. J. Salmond, and A. F. M. Smith, “Novel approach
to nonlinear/non-Gaussian Bayesian state estimation,” Proc. Inst. Elect.
Eng. F, vol. 140, no. 2, pp. 107–113, Apr. 1993.
[7] H. J. Kushner, “Approximations to optimal nonlinear filters,” IEEE
Trans. Automat. Contr., vol. AC-12, pp. 546–556, Oct. 1967.
[8] H. W. Sorenson and A. R. Stubberud, “Non-linear filtering by approximation of the a posteriori density,” Int. J. Contr., vol. 8, no. 1, pp. 33–51,
1968.
[9] F. E. Daum, “New exact nonlinear filters,” in Bayesian Analysis of
Time Series and Dynamic Models, J. C. Spall, Ed. New York: Marcel
Dekker, 1988, pp. 199–226.
[10] S. J. Julier, J. K. Uhlmann, and H. F. Durrant-Whyte, “A new approach
for filtering nonlinear systems,” in Proc. Am. Contr. Conf., Seattle, WA,
1995, pp. 1628–1632.
[11] S. J. Julier and J. K. Uhlmann. (1994, Aug.) A general method for approximating nonlinear transformations of probability distributions. [Online]. Available: http://www.robots.ox.ac.uk/~siju
[12]
, “A new extension of the Kalman filter to nonlinear systems,”
in Proc. AeroSense: 11th Int. Symp. Aerosp./Defense Sensing, Simulat.
Contr., Orlando, FL, 1997.
4This can be contrasted with an alternative approach of the initial intuition,
which was explored in [26]. Under that scheme, no copies of the previously
estimated mean are included in the sample set and the sigma points are scaled
using a parameter . In the limit, as tends to infinity, this algorithm predicts
the same mean and covariance as the EKF. However, when 1; this method
estimates the same mean and covariance as the new filter, but with = 0:
=
[13] P. S. Maybeck, Stochastic Models, Estimation, and Control. New
York: Acdemic, 1979, vol. 1.
[14] M. Athans, R. P. Wishner, and A. Bertolini, “Suboptimal state estimation for continuous-time nonlinear systems from discrete noisy measurements,” IEEE Trans. Automat. Contr., vol. 13, pp. 504–518, Oct. 1968.
[15] D. E. Catlin, “Estimation, control and the discrete Kalman filter,” in
Applied Mathematical Sciences 71. New York: Springer-Verlag, 1989,
p. 84.
[16] S. Clark, “Autonomous land vehicle navigation using millimetre wave
radar,” Ph.D. dissertation, University of Sydney, Australia, 1999.
[17] D. Lerro and Y. K. Bar-Shalom, “Tracking with debiased consistent converted measurements vs. EKF,” IEEE Trans. Aerosp. Electron. Syst., vol.
29, pp. 1015–1022, July 1993.
[18] J. K. Uhlmann, “Simultaneous map building and localization for real
time applications,” Univ. Oxford, U.K., Tech. Rep., 1994.
[19] S. J. Julier and J. K. Uhlmann, “A consistent, debiased method for
converting between polar and Cartesian coordinate systems,” in Proc.
AeroSense: 11th Int. Symp. Aerosp./Defense Sensing, Simulat. Contr.,
Orlando, FL, 1997.
[20] S. J. Julier, “A skewed approach to filtering,” in Proc. AeroSense: 12th
Int. Symp. Aerosp./Defense Sensing, Simulat. Contr., Orlando, FL, 1998.
, “Comprehensive process models for high-speed navigation,”
[21]
Ph.D. dissertation, Univ. Oxford, U.K., 1997.
[22] A. Montobbio, “Sperimentazione ed affinamento di un localizzatore,”
B.S. thesis, Politechnico di Torino, Italy, 1998.
[23] R. Smith, “Navigation of an underwater remote operated vehicle,” Univ.
Oxford, Tech. Rep., 1995.
[24] T. S. Schei, “A finite difference approach to linearization in nonlinear
estimation algorithms,” in Proc. Am. Contr. Conf., vol. 1, Seattle, WA,
1995, pp. 114–118.
[25] R. L. Bellaire, E. W. Kamen, and S. M. Zabin, “A new nonlinear iterated
filter with applications to target tracking,” in Proc. AeroSense: 8th Int.
Symp. Aerosp./Defense Sensing, Simulat. Contr., vol. 2561, Orlando, FL,
1995, pp. 240–251.
[26] B. M. Quine, J. K. Uhlmann, and H. F. Durrant-Whyte, “Implicit Jacobians for linearized state estimation in nonlinear systems,” in Proc.
Amer. Contr. Conf., Seattle, WA, 1996, pp. 1645–1646.
An
Boundary Layer in Sliding Mode for
Sampled-Data Systems
Wu-Chung Su, Sergey V. Drakunov, and Ümit Özgüner
Abstract—The use of a discontinuous control law (typically, sign functions) in a sampled-data system will bring about chattering phenomenon
in the vicinity of the sliding manifold, leading to a boundary layer with
is the sampling period. However, by proper
thickness ( ) where
consideration of the sampling phenomenon in the discrete-time sliding
mode control design, the thickness of the boundary layer can be reduced
to ( ) In contrast to discontinuous control for continuous-time VSS,
the discrete-time sliding mode control need not be of switching type.
Index Terms—Boundary layer, discrete-time sliding mode, sampled-data
systems.
Manuscript received April 3, 1998; revised August 31, 1998, and February
28, 1999. Recommended by Associate Editor, S. Hara. This work was
supported by the National Science Council of Taiwan, under Project NSC
89-2213-E-005-026.
W.-C. Su is with the Department of Electrical Engineering, National ChungHsing University, Taichung, Taiwan (e-mail: [email protected]).
S. V. Drakunov is with the Department of Electrical and Computer Engineering, Tulane University, New Orleans, LA 70118 USA.
Ü. Özgüner is with the Department of Electrical Engineering, The Ohio State
University, Columbus, OH 43210 USA.
Publisher Item Identifier S 0018-9286(00)02146-2.
0018–9286/00$10.00 © 2000 IEEE