Anda di halaman 1dari 11

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO.

9, SEPTEMBER 2007 1631

On Unscented Kalman Filtering for State Estimation


of Continuous-Time Nonlinear Systems
Simo Särkkä, Member, IEEE

Abstract—This paper considers the application of the unscented mean and covariance of the posterior state process. A square
Kalman filter (UKF) to continuous-time filtering problems, root version of the filter is also derived and it consists of a
where both the state and measurement processes are modeled matrix differential equation for the sigma points.
as stochastic differential equations. The mean and covariance
differential equations which result in the continuous-time limit of The continuous-discrete UKF is derived as a special case of
the UKF are derived. The continuous-discrete UKF is derived as a the continuous-time filter, when the continuous-time prediction
special case of the continuous-time filter, when the continuous-time equations are combined with the update step of the discrete-time
prediction equations are combined with the update step of the UKF. A square root version of this continuous-discrete filter is
discrete-time UKF. The filter equations are also transformed into also presented.
sigma-point differential equations, which can be interpreted as
matrix square root versions of the filter equations.
II. PROBLEM STATEMENT
Index Terms—Continuous-discrete filter, continuous-time filter,
continuous-time state space model, nonlinear state space model,
nonlinear system, stochastic differential equation, unscented A. Optimal Continuous-Time Filtering
Kalman filter (UKF). In analog communication systems [11] the measured signals
are typically continuous-time processes and analog receivers are
I. INTRODUCTION devices, which demodulate or estimate the actual transmitted
continuous-time signals from the noisy measured signals. Also
in many analog (electrical or mechanical) control systems (see,

T HE unscented Kalman filter (UKF) [1]–[3] is an efficient


derivative free filtering algorithm for computing approxi-
mate solutions to discrete-time nonlinear optimal filtering prob-
e.g., [13] and [14]) operating without digital computers the mea-
sured signals are continuous-time, not discrete-time signals. The
optimal recursive estimation of this kind of systems is called op-
lems. It has been successfully applied to numerous practical timal continuous-time filtering [16], [17].
problems and it has been shown to outperform the extended The most general form of the optimal continuous-time fil-
Kalman filter (EKF) in many cases [4]. However, in its original tering models considered in this paper is
form, the UKF is a discrete-time algorithm and it cannot be di-
rectly applied to continuous-discrete filtering problems, where
the state dynamics are modeled as continuous-time stochastic (1)
processes, or to continuous-time filtering problems, where both
the state and measurement processes are modeled as contin- where is the state process; is the
uous-time stochastic processes. (integrated) measurement process; is the drift function; is
Phenomena, which can be modeled as time varying systems, the measurement model function; and are arbitrary
where a continuous-time signal is observed discretely or contin- time varying matrices, independent of and ; and
uously in time are very common in engineering and physics ap- are independent Brownian motions with diagonal diffu-
plications. Examples of such applications are GPS and inertial sion matrices and , respectively.The dynamic and
navigation [5], [6], target tracking [7]–[9], estimation of biolog- measurement models can be equivalently interpreted as Ito or
ical processes [10], telecommunications [11], [12], stochastic Stratonovich type stochastic differential equations [19], [20].
optimal control [13], [14], and inverse problems in physics [15]. The filtering model can also be formulated in terms of formal
In this paper, the differential equations which result in the white noises , , and differen-
continuous-time limit of the UKF are derived. Both the contin- tial measurement as follows [16]:
uous-time and continuous-discrete cases are considered. The
derived continuous-time filtering equations, which could be
called the unscented Kalman-Bucy filter (UKBF) equations, are
similar to the extended Kalman-Bucy filter (EKBF) equations (2)
[16]–[18] and consist of a pair of differential equations for the
where the white noise processes and have spectral den-
sities and , respectively.
Manuscript received May 8, 2006; revised January 24, 2007. Recommended The purpose of the optimal (Bayesian) continuous-time fil-
by Associate Editor A. Garulli.
The author is with the Helsinki University of Technology, FIN-02015 HUT, tering is to recursively compute the posterior distribution
Finland (e-mail: simo.sarkka@hut.fi).
Digital Object Identifier 10.1109/TAC.2007.904453 (3)
0018-9286/$25.00 © 2007 IEEE
1632 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO. 9, SEPTEMBER 2007

or at least the relevant moments of the distribution (e.g., mean where is the time of measurement . By using optimal pre-
and covariance). The formal solution to the filtering problem is diction the corresponding distribution can also be computed for
well known and it is given by the Kushner-Stratonovich equa- all time instances before the next measurement .
tion [21]–[23], which is a measure valued stochastic partial dif- In theory, the solution to the continuous-discrete filtering
ferential equation. The unnormalized version of the equation is problem can be computed by the following prediction and
called the Zakai equation [24], [25]. However, these equations update steps [16]:
only give the formal solution, and the actual computation of the • Prediction step: solves the predicted probability density at
distribution or its expectations (such as mean and covariance) time step from the Kolmogorov forward partial differ-
would require an infinite amount of computational resources. In ential equation using the old posterior probability density
certain special cases, the equations do have finite dimensional at time step as the boundary condition.
solutions, which leads to Kalman-Bucy filters [26] and Benes • Update step: uses the Bayes’ rule for computing the pos-
filters [27]. terior probability density of state at time step from the
Because the exact solution to the continuous-time optimal fil- predicted probability density of the prediction step, and the
tering problem is generally intractable, approximations must be likelihood of the measurement .
used. The most common approximation method is the EKBF As in the continuous-time case, the closed form solutions to
(see, e.g., [16]–[18]), which approximates the exact solution by these equations or to the equations for the moments (e.g., mean
replacing the nonlinear model with a suitably linearized approx- and covariance) of the distributions can only be found in a
imate model, which can be solved by the Kalman-Bucy filter. few special cases [30]–[32] and approximations are generally
Another general way of forming approximations is Monte Carlo needed. The continuous-discrete EKF [16]–[18] uses a Taylor
sampling [28], [29], where a set of weighted particles is used for series expansion approximation to the nonlinear drift function
approximating the posterior probability measure. and forms a Gaussian process approximation to the SDE.
Another possible approach is to simulate sample paths of SDEs
B. Optimal Continuous-Discrete Filtering [33] and use particle filters for estimation [34], [35]. Interacting
and branching particle systems [36] are particle-based solutions
Nowadays, in many signal processing systems, the sensor
to nonlinear filtering problems also in the continuous-discrete
measurements are obtained at discrete instances of time either
setting.
due to sampling, due to processing delays in the device or be-
Other possible approaches are statistical linearization [17],
cause the sensor operates in scans. Still in the Nature time is
grid-based methods [35], [37] and multiple model methods [7],
continuous, not discrete, and for this reason often a physically
[35], Gaussian sum approximations [38], [39], and numerical
more realistic approach than discrete-time filtering or contin-
solving of the Kolmogorov forward equation [40], [41].
uous-time filtering is continuous-discrete filtering [16], [17]. In
continuous-discrete filtering the state dynamics are modeled as III. DISCRETE-TIME UNSCENTED KALMAN FILTERING
continuous-time stochastic processes, and the measurements are
obtained at discrete instances of time. This differs from dis- A. Unscented Transform
crete-time filtering, because in that approach both the dynamics The unscented transform (UT) [1]–[3] can be used for
and measurements are modeled as discrete-time processes. forming a Gaussian approximation to the joint distribution
The continuous-discrete filtering models considered here of random variables and , when the random variable is
have the general form obtained by a nonlinear transformation of the Gaussian random
variable as follows:

(4)
(7)
where is the state; is the measurement; where , , and is a general non-
is the drift function; is the dispersion matrix; is linear function. The idea of UT is to form a fixed number of
Brownian motion with diffusion matrix; , is the mea- deterministically chosen sigma-points, which capture the mean
surement model function; is a zero mean Gaussian measure- and covariance of the original distribution of exactly. These
ment noise with covariance matrix . The dynamic model can sigma-points are then propagated through the nonlinearity and
be equivalently interpreted as a Ito or Stratonovich type sto- the mean and covariance of the transformed variable are esti-
chastic differential equation. As in the continuous-time filtering mated from them. Note that the unscented transform is signifi-
case the dynamic model can also be written in terms of white cantly different from Monte Carlo estimation, because the sigma
noise process as points are selected deterministically [1], [2].
Algorithm 3.1 (Unscented Transform): The unscented trans-
(5) form can be used for forming the Gaussian approximation

The purpose of (Bayesian) continuous-discrete filtering is to re- (8)


cursively compute the posterior distribution
to the joint probability density of and . The
(6) unscented transform is the following:
SÄRKKÄ: UKF FOR STATE ESTIMATION 1633

1) Form the set of sigma points from the columns of B. UKF


the matrix as follows:
The UKF [1]–[3] is a discrete-time filtering algorithm, which
utilizes the unscented transform for computing approximate so-
lutions to the filtering problems of the form1

(9)
(18)
and compute the associated weights
where is the state, is the measurement,
is a Gaussian process noise ,
and is a Gaussian measurement noise .
The mean and covariance of the initial state are and ,
respectively.
In terms of the unscented transform the UKF predic-
tion and update steps can be written as follows.
• Prediction: Compute the predicted state mean and the
(10)
predicted covariance as

The parameter is a scaling parameter defined as

(11) (19)

The positive constants , and are used as parameters • Update: Compute the predicted mean and covariance of
of the method. the measurement , and the cross-covariance of the state
2) Transform each of the sigma points as and measurement :

(12)

3) Mean and covariance estimates for can be computed as (20)

Then compute the filter gain , the state mean and


the covariance , conditional to the measurement

(13)

4) The cross-covariance of and can estimated as (21)

The filtering is started from the initial mean and covariance


(14) . A thorough treatment of the unscented Kalman filtering,
sigma-point filtering in general and connections to several other
The matrix square root of positive definite matrix means a filtering algorithms can be found in [4]. Efficient square root
matrix such that versions of the UKF are presented in [4] and [43].

(15)
IV. CONTINUOUS-TIME UNSCENTED KALMAN FILTERING
Because the only requirement for is the definition above, we
can, for example, use the lower triangular matrix of the Cholesky A. Matrix Form of UKF
factorization (see, e.g., [42]). In this section, the UKF is presented in matrix form, where
In this paper, the Cholesky factorization is denoted as the the weighted sums of sigma points are written as equivalent ma-
function trix expressions. This eases the derivations of continuous-time
forms of the UKF in Section IV-B.
(16)
In order to clean up the notation, we shall use the convention
The unscented transform can be seen as a function (or func- that
tional) from to
(22)
1In this paper, we only consider the case of additive noise, but UKF can also
(17) be applied to more general filtering problems with nonadditive noise.
1634 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO. 9, SEPTEMBER 2007

where and means that the th column B. Continuous-Time UKF


of the matrix is formed as follows:
By taking the formal limit of the discrete-time UKF equations
in Algorithm 4.1, the following novel continuous-time filter can
be derived.
(23)
Algorithm 4.2 (UKBF): The stochastic differential equations
corresponding to the UKF in the continuous-time limit of the
Lemma 4.1 (The Matrix Form of UT): The unscented trans-
state and measurement processes, that is, the UKBF equations,
form can be written in matrix form as follows:
are given as

(24)

where is the matrix of sigma points, , and (29)


vector and matrix are defined as follows:
where we have formally defined the differential measurement
. The sigma-point matrix is defined as

(30)
(25) Proof: See Appendix I-B.
To avoid problems related to the finite numerical precision of
Proof: See Appendix I-A. computer arithmetic, Kalman filter equations are often imple-
Algorithm 4.1 (UKF): The UKF prediction and update steps mented such that the matrix square roots of covariance matrices
can be written in matrix form as follows. are used in computations instead of their actual values [18]. Be-
• Prediction: Compute the predicted state mean and the cause UKF already uses matrix square roots in its sigma-points,
predicted covariance as the square root version of continuous-time UKF can be obtained
by formulating the filter as a differential equation for the sigma
points.
Algorithm 4.3 (Square Root UKBF): The UKBF can be for-
mulated in terms of sigma points as follows2:

(26)

• Update: Compute the predicted mean and covariance of


the measurement , and the cross-covariance of the state
and measurement

(31)

where the matrix of sigma-points is defined as

(32)
(27)
and is a function returning the lower diagonal part of the
Then compute the filter gain , the state mean and
argument as follows:
the covariance
if
if (33)
if .
(28) 2Here () denotes the ith column of the argument matrix.
SÄRKKÄ: UKF FOR STATE ESTIMATION 1635

The initial should be the lower triangular Cholesky factor Proof: Formally set in Algorithm 4.3, which
of the initial covariance matrix . results in .
Proof: See Appendix I-D.
Note that in the square root filter, the equations contain the D. Continuous-Discrete UKF
matrix , which is the lower triangular Cholesky factor of
the covariance at all time instances . The mean and covariance form of continuous-discrete UKF
However, the factorization needs to be explicitly computed only can be now implemented as follows.
on the initial time step, because can be extracted from Algorithm 4.6 (Continuous-Discrete UKF): The prediction
at any time step by very simple means. This also implies that in and update steps of the continuous-discrete UKF are the fol-
the implementation of the algorithm the covariance of the lowing.
state never needs to be evaluated. • Prediction. Integrate the differential equations(34) in Al-
The UKBF equations presented in this section are actually gorithm 4.4 from the initial conditions ,
Stratonovich type of stochastic differential equations, because to time instance . The predicted mean
they have been derived by taking the continuous-time limit of and covariance are given as and
discrete-time equations by using the rules of conventional cal- , respectively.
culus. In the beginning of the paper it was stated that the con- • Update. The update step is the same as the discrete-time
tinuous-time dynamic and measurement models can be equiva- UKF update step (27) of Algorithm 4.1.
lently interpreted as either Ito or Stratonovich type of stochastic The corresponding square root continuous-discrete UKF has
differential equations, which is indeed true, but with both the the same form.
interpretations the filter equations still are of the Stratonovich Algorithm 4.7 (Square Root Continuous-Discrete UKF): The
type. prediction and update steps of the square root continuous-dis-
crete UKF are the following:
C. Unscented Continuous-Time Prediction • Prediction. Integrate the differential equations(35) in Al-
In this section the continuous-time prediction equations are gorithm 4.5 from the initial conditions , which are
derived as special cases of the continuous-time filters in the pre- the sigma points generated from and
vious section. The motivation is that using the prediction equa- . The integration is continued up to the
tions we can construct the continuous-discrete versions of the time instance and the predicted sigma points are given
UKF. as .
Algorithm 4.4 (Mean and Covariance Prediction): The pre- • Update. The update step is the same as the discrete-time
dicted mean and covariance of the state for times UKF update step (27) of Algorithm 4.1, but now the
given the mean and covariance at the time instance , predicted sigma points do not have to be generated from
that is, and can be computed by integrating the dif- mean and covariance, because they are already available.
ferential equations To enhance the numerical stability of the update equations,
square root versions [4], [43] of the UKF update step can
also be used.

V. REMARKS AND DISCUSSION


(34)
A. Computational Complexity
from initial conditions and to time instance . Here
is defined as in (30), Although, the new matrix form of the unscented transform is
Proof: Formally set in Algorithm 4.2, which very useful in derivation of the filtering equations, in numerical
results in . computations, the classical form can be computationally more
Algorithm 4.5 (Square Root Prediction): The UKF prediction efficient. The matrix expressions appearing in the filtering equa-
equations can be written in terms of sigma points as tions can be equivalently written as summations as follows:

(35)

where , , and are defined as in Algorithm 4.3.


The integration is started from the sigma points , which (36)
are generated from and .
1636 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO. 9, SEPTEMBER 2007

where TABLE I
RMSE VALUES OF THE PARAMETER AND SIGNAL AVERAGED OVER 1000
MONTE CARLO RUNS IN THE CONTINUOUS-TIME NONLINEAR
FILTERING PROBLEM

(37)

By using the summation forms of the covariance terms the full


matrix product is replaced by the weighted sum of outer prod-
ucts, which is lighter to compute.
The relationship between the UKBF and EKBF approxima-
tions can be seen to be

(38)

where the elements of the Jacobian matrices and


are given as

(39)
Fig. 1. Illustration of transient period error behavior in continuous-time non-
linear filtering problem.
The computational complexity of the UKBF (or continuous-dis-
crete UKF) can be seen to be two–three times the computational
complexity of the EKBF, when compared in terms of number of models are in continuous time. The simulation was performed
multiplications and additions. When the state dimension is , over time period of 5 s using the Euler-Maruyama scheme [33]
the UKBF needs evaluations of and , when EKBF and with time steps of .
needs only one. However, in addition to that, EKBF needs eval- The root mean squared error (RMSE) results of 1000 Monte
uations of the Jacobian matrices (and possibly Hessian matrices) Carlo runs with EKBF, second-order extended Kalman-Bucy
of both the functions. filter (EKBF2), UKBF, and statistically linearized filter (SLF)
are shown in Table I. It can be seen that SLF is the best of the
VI. ILLUSTRATIVE EXAMPLES filters in RMSE sense, because it uses the closed form formulas
for expectations and covariances of the nonlinearities (see [17]),
A. Continuous-Time Nonlinear Filtering when the other filters can only approximate them. The UKBF
In this section, we shall test the performance of the UKBF gives a significantly better result than EKBF and EKBF2, and
in continuous-time version of the nonlinear filtering problem, the UKBF result is also quite close to the base line result of SLF.
which was used in [17] for demonstrating the performance of the The performances of EKBF and EKBF2 are most likely that bad
statistical linearization-based filter. Both the state and the because the Taylor series expansion-based approximations do
measurements are one-dimensional continuous-time pro- not work well when the estimation error is large.
cesses and the filtering model is Even thought the SLF is better than UKBF in RMSE sense,
SLF has the serious disadvantage that in order to implement it,
one has to be able to compute closed form formulas for cer-
tain expected values of nonlinear transformations of Gaussian
(40) random variables [17]. These expected values can be computed
in closed form only in simple special cases and, thus, UKBF is a
where and are continuous-time white noise processes very good choice for models with significant nonlinearities and
with spectral densities and , respectively. uncertainties.
Note that in [17] the dynamic model was in continuous time and A typical transient behavior of the filters is illustrated in
the measurement model was in discrete time, but here both the Fig. 1. It can be seen that SLF is best because it converges very
SÄRKKÄ: UKF FOR STATE ESTIMATION 1637

quickly near the correct state. The convergence of UKBF is modified discrete process noise covariance is denoted as
slower, but significantly faster than of EKBF and EKBF2. and the corresponding spectral density is denoted as .
The radar is located at and the measure-
B. Reentry Vehicle Tracking ment model is
Here we consider the reentry tracking problem, where a radar
is used for tracking a space vehicle, which enters the atmosphere
at a very high speed. The reentry problem was used for demon-
strating the performance of UKF in [2] and slight corrections to (44)
the equations and simulation parameters were later published in
[44]. The stochastic equations of motion for the space vehicle where the and . In the simula-
are given as [2], [44] tion, the assumed means and covariances of the state, simulated
means and covariances of the state and the standard deviations
of measurements were selected to be the same as in [2], [44].
The simulated data were generated by simulating the stochastic
differential (41) with 100 steps of Euler-Maruyama scheme [33]
between each measurement.
The discrete UKF (DUKF) and continuous-discrete
UKF (CDUKF) were both implemented using 10 steps of
Runge-Kutta integration between measurements, but the differ-
ence was in handling of process noise.
• In DUKF, each measurement is processed as follows.
1) Integrate each of the sigma point through the noise free
dynamic model using 10 steps of the Runge-Kutta in-
tegration.
2) Compute the predicted mean and covariance, and
model process noise effect by approximating the dis-
crete covariance by .
3) Perform standard UKF update step for the measure-
ment.
(41)
• In CDUKF, each measurement is processed as follows.
where , , are white Gaussian process noises 1) Integrate the mean and covariance differential equa-
with known joint spectral density. The constants are [2] tions using 10 steps of the Runge-Kutta integration,
and using as the diffusion matrix (or spectral den-
sity) of the process noise.
2) Perform standard UKF update step for the measure-
ment.
The amount of computations required by the CDUKF is slightly
(42) higher than of DUKF, but the number of evaluations of the dy-
In [2], the radar measured 10 times per second and the simulated namic model function is the same for both the models and thus
discrete process noise covariance was the practical total difference is small. The algorithm parameters
in the unscented transforms were selected to be , ,
.
(43)
Simulations were performed using different time steps of
seconds and the results from 100 Monte
If we interpret the dynamic model (41) as originally contin- Carlo simulations per step size are shown in Fig. 2. The DUKF
uous time model, we may assume that the discretized covari- and CDUKF have very much the same performance when the
ance is actually an approximation to a nonlinear continuous- time step size is short. However, when the time step grows,
time process driven by continuous-time process noise with spec- DUKF encounters numerical problems and its error grows
tral density . In this interpretation, it is reasonable to assume rapidly. The problem is that when the sigma-points are inte-
that the relationship between the discrete covariance and the grated one at a time, nothing prevents the covariance estimate
continuous-time spectral density is originally the approximation from becoming nonpositive definite. With larger time steps
where is the sampling time used in this causes severe numerical difficulties. At the same time, no
[2]. Thus the true spectral density matrix is the matrix in (43) numerical problems can be seen in the CDUKF and its error
multiplied by 10. grows much slower.
In the actual filter implementation, the lower right corner term In this simulation scenario, the advantage of the continuous-
in the modeled process noise covariance was set to the value time formulation over the discrete-time formulation is the nu-
. This additional small noise term was used for the constant merical stability. This stability is due to that when the noise
parameter to enhance the filter stability as was done in [44]. This process is modeled as a continuous-time process the nonlinear
1638 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO. 9, SEPTEMBER 2007

then the sigma point computation in (9) can be written in form


of the first equation in (24). The second equation is simply the
vector form of (12).
If we define the weight vector and matrix as in (25)
and denote the matrix of sigma points of as the transformed
mean and covariance equations can be written as

(46)

Fig. 2. RMSE versus time step size in the reentry filtering problem. Results
are from 100 Monte Carlo runs per time step with the CDUKF and DUKF. (47)
The higher errors of DUKF with longer time step sizes are caused by numerical
problems on the prediction step.

dynamic model cannot force the covariance to become nonposi-


tive definite. In the discrete-time formulation the covariance can
become nonpositive definite.
In this particular problem, it is not essential whether uncer-
tainties are modeled as discrete-time or continuous-time sto-
chastic processes. Since the performance of CDUKF seems to (48)
be at least that of DUKF, in cases where the continuous-time sto-
chastic process formulation is more accurate in modeling point which leads to last three equations in (24).
of view, the CDUKF is likely to perform better than the discrete Derivation of UKBF: The continuous-time UKF equations
UKF. can be derived from the discrete-time UKF by the same limiting
procedure as is commonly used in derivation of the Kalman-
Bucy filter from the discrete-time Kalman filter (see, e.g., [14],
VII. CONCLUSION [18]). The continuous-time filtering model

In this paper, novel continuous-time and continuous-discrete


versions of the originally discrete-time UKF have been derived (49)
and applied to nonlinear continuous-time filtering and reentry
vehicle tracking problems. Numerically more stable square-root where and are independent Brownian motions with
versions of the new filters have also been derived. diagonal diffusion matrices and can be interpreted
The continuous-time and continuous-discrete unscented fil- such that when is sufficiently close to zero, we have
ters are good alternatives to the EKFs in models, where the Ja-
cobian and Hessian matrices of the drift terms are not available (50)
or when the Taylor series expansion approximations do not work
well. According to the simulations, the approximations gener- where , , and
ated by the new filters are better in the cases where the model is a function such that when .
nonlinearities and estimation uncertainties are significant. Assuming that , the UKF prediction
from to can be written as

APPENDIX
DERIVATIONS

Derivation of Matrix Form of UT: If we define the matrix


of sigma points as

(45) (51)
SÄRKKÄ: UKF FOR STATE ESTIMATION 1639

Substituting into the equations of and Alternative Derivation: In Appendix I-B, the equations
, and using the identities , of the UKBF were derived by taking the formal limit of the
gives discrete-time equations. However, it is also possible to derive
the same equations purely in continuous-time framework. The
mean and covariance of the optimal continuous-time filter are,
in theory, given by the stochastic differential equations [16]

(52)

Assuming that we measure the difference


the UKF update step can be written as

(56)
The expectations are with respect to the posterior distribution
(53)
of the state and thus cannot be in practice computed, be-
Substituting , , , , cause the computation would require the knowledge of all (in-
and retaining only first-order terms results in finite number of) moments of the distribution. However, an un-
scented transform-based approximation to these equations can
be formed as follows. If we assume that the third-order term in
the covariance equation above is approximately zero, the equa-
tions above can be approximately written as

(54)
(57)
Substituting and gives where denotes the cross-co-
variance of and , and is the differentiated
measurement process. If we assume that the posterior mean and
covariance of are and , respectively, unscented
transform-based approximations to the expectations and covari-
ances can be formed as follows:

(55)

Dividing by and taking the limit results in the


differential equations in the Algorithm 4.2. (58)
1640 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 52, NO. 9, SEPTEMBER 2007

where are the sigma-points generated from and . Taking derivatives from both sides of the equation gives
Substituting these approximations into the (57) results in the
UKBF equations in Algorithm 4.2.
Note that the above derivation is not strictly complete as such,
(66)
because the UKBF equations presented in Section IV-B are ac-
Substituting the equations for and leads to
tually Stratonovich equations, but the (57) are Ito type of equa-
equations in Algorithm 4.3.
tions. However, we may argue that the extra terms arising from
the conversion of (57) from Ito from to Stratonovich form would
only introduce terms involving odd moments, which anyway
disappear due to the Gaussian approximation. ACKNOWLEDGMENT
Derivation of SR-UKBF: Assume that the matrix square
The author would like to thank A. Vehtari, J. Lampinen, I.
root used is the Cholesky factorization
Kalliomäki, H. Gruschinski, and the anonymous reviewers for
helpful comments.
(59)

which also assures that the square root is lower triangular. If we


REFERENCES
define (symmetric matrix)
[1] S. J. Julier, J. K. Uhlmann, and H. F. Durrant-Whyte, “A new method
for the nonlinear transformation of means and covariances in filters and
estimators,” IEEE Trans. Autom. Control, vol. 45, pp. 477–482, Mar.
2000.
[2] S. J. Julier and J. K. Uhlmann, “Unscented filtering and nonlinear esti-
mation,” Proc. IEEE, vol. 92, pp. 401–422, Mar. 2004.
[3] E. A. Wan and R. M. van der, “The unscented Kalman filter,” in Kalman
(60) Filtering and Neural Networks, S. Haykin, Ed. New York: Wiley,
2001, ch. 7.
[4] R. M. van der, “Sigma-point Kalman filters for probabilistic inference
then the differential equation of the covariance is of the form in dynamic state-space models,” Ph.D., OGI School of Sci. Eng.,
Oregon Health and Sci. Univ., Portland, OR, 2004.
[5] E. D. Kaplan, Understanding GPS, Principles and Applica-
(61) tions. Boston, MA: Artech House, 1996.
[6] M. S. Grewal, L. R. Weill, and A. P. Andrews, Global Positioning Sys-
tems, Inertial Navigation and Integration. New York: Wiley Intersci.,
It is now possible to derive differential equations for the 2001.
Cholesky factor using the procedure presented in [45] and [7] Y. Bar-Shalom, X.-R. Li, and T. Kirubarajan, Estimation With Applica-
later in [14]. We first expand the derivative of the covariance tions to Tracking and Navigation. New York: Wiley Intersci., 2001.
[8] S. Blackman and R. Popoli, Design and Analysis of Modern Tracking
matrix as follows: Systems. Boston, MA: Artech House Radar Library, 1999.
[9] L. D. Stone, C. A. Barlow, and T. L. Corwin, Bayesian Multiple Target
Tracking. Boston, MA: Artech House, 1999.
(62) [10] J. D. Murray, Mathematical Biology. New York: Springer, 1993, vol.
19.
[11] H. L. Van Trees, Detection, Estimation, and Modulation Theory Part
Multiplying both sides from left by and from right by I. New York: Wiley, 1968.
[12] H. L. Van Trees, Detection, Estimation, and Modulation Theory Part
gives II. New York: Wiley, 1971.
[13] P. Maybeck, Stochastic Models, Estimation and Control. New York:
Academic, 1982, vol. 3.
[14] R. F. Stengel, Optimal Control and Estimation. New York: Dover,
1994.
(63) [15] J. Kaipio and E. Somersalo, Statistical and Computational Inverse
Problems, ser. Applied Mathematical Sciences. New York: Springer,
2005.
Now the left-hand side (LHS) is sum of lower triangular [16] A. H. Jazwinski, Stochastic Processes and Filtering Theory. New
matrix and upper triangular matrix York: Academic, 1970.
[17] A. Gelb, Ed., Applied Optimal Estimation. Cambridge: The MIT
, and the right-hand side (RHS) is sym- Press, 1974.
metric matrix. Thus it can be concluded that the derivative of [18] M. S. Grewal and A. P. Andrews, Kalman Filtering, Theory and Prac-
can be written as tice Using MATLAB. New York: Wiley Intersci., 2001.
[19] I. Karatzas and S. E. Shreve, Brownian Motion and Stochastic Cal-
culus. New York: Springer, 1991.
[20] B. Øksendal, Stochastic Differential Equations: An Introduction with
(64) Applications. New York: Springer, 2003.
[21] H. J. Kushner, “On the differential equations satisfied by conditional
where the function , which is defined in (33) gives the lower probability densities of markov processes, with applications,” J. SIAM
Control Ser. A, vol. 2, no. 1, 1964.
diagonal part and half of the diagonal part of its argument. The [22] R. L. Stratonovich, Conditional Markov Processes and Their Applica-
expression of the sigma points is tion to the Theory of Optimal Control. New York: American Elsevier,
1968.
[23] R. S. Bucy, “Nonlinear filtering theory,” IEEE Trans. Autom. Control,
(65) vol. 10, pp. 198–198, 1965.
SÄRKKÄ: UKF FOR STATE ESTIMATION 1641

[24] M. Zakai, “On the optimal filtering of diffusion processes,” Zeit. [38] D. L. Alspach and H. W. Sorenson, “Nonlinear Bayesian estimation
Wahrsch, vol. 11, pp. 230–243, 1969. using Gaussian sum approximations,” IEEE Trans. Autom. Control,
[25] G. Kallianpur, Stochastic Filtering Theory. New York: Springer- vol. AC-17, Aug. 1972.
Verlag, 1980. [39] T. Söderström, Discrete-Time Stochastic Systems: Estimation and Con-
[26] R. E. Kalman and R. S. Bucy, “New results in linear filtering and pre- trol. Englewood Cliffs, NJ: Prentice-Hall, 1994.
diction theory,” Trans. ASME, J. Basic Eng., vol. 83, pp. 95–108, Mar. [40] J. Gunther, R. Beard, J. Wilson, T. Oliphant, and W. Stirling, “Fast non-
1961. linear filtering via galerkin’s method,” in Amer. Control Conf. Amer.
[27] V. E. Benes, “Exact finite-dimensional filters for certain diffusions with Autom. Control Council, 1997.
nonlinear drift,” Stochastics, vol. 5, pp. 65–92, 1981. [41] S. Challa, Y. Bar-Shalom, and V. Krishnamurthy, “Nonlinear filtering
[28] D. Crisan and T. Lyons, “Nonlinear filtering and measure-valued pro- via generalized edgeworth series and Gauss-Hermite quadrature,”
cesses,” Probab. Theory Relat. Fields, vol. 109, pp. 217–244, 1997. IEEE Trans. Signal Process., vol. 48, Jun. 2000.
[29] D. Crisan and T. Lyons, “A particle approximation of the solution of [42] G. H. Golub and C. F. van Loan, Matrix Computations, 3rd ed. Bal-
the kushner-stratonovitch equation,” Probab. Theory Relat. Fields, vol. timore, MD : The Johns Hopkins Univ. Press, 1996.
115, pp. 549–578, 1999. [43] R. M. van der and E. Wan, “The square-root unscented Kalman filter
[30] R. E. Kalman, “A new approach to linear filtering and prediction prob- for state and parameter-estimation,” in IEEE Int. Conf. Acoust., Speech,
lems,” Trans. ASME, J. Basic Eng., vol. 82, pp. 35–45, Mar. 1960. Signal Process. ICASSP, 2001, vol. 6, pp. 3461–3464.
[31] F. E. Daum, “Exact finite-dimensional nonlinear filters for continuous [44] S. J. Julier and J. K. Uhlmann, “Corrections to unscented filtering and
time processes with discrete time measurements,” in Proc. 23rd Conf. nonlinear estimation,” Proc. IEEE, vol. 92, pp. 1958–1958, Dec. 2004.
Decision Control, 1984, pp. 16–22. [45] M. Morf, B. Lévy, and T. Kailath, “Square-root algorithms for the
[32] F. E. Daum, “Exact finite-dimensional nonlinear filters,” IEEE Trans. continuous-time linear least-square estimation problem,” IEEE Trans.
Autom. Control, vol. AC-31, pp. 616–622, 1986. Autom. Control, vol. 23, pp. 907–911, Oct. 1978.
[33] P. E. Kloeden and E. Platen, Numerical Solution to Stochastic Differ-
ential Equations. New York: Springer, 1999.
[34] A. Doucet, N. de Freitas, and N. Gordon, Eds., Sequential Monte Carlo
Methods in Practice. New York: Springer, 2001. Simo Särkkä (M’06) received the M.Sc. degree in
[35] B. Ristic, S. Arulampalam, and N. Gordon, Beyond the Kalman physics and the D.Sc. degree in electrical engineering
Filter. Boston, MA: Artech House, 2004. from Helsinki University of Technology, Espoo, Fin-
[36] P. D. Moral and L. Miclo, “Branching and interacting particle sys- land, in 2000 and 2006, respectively.
tems. Approximations of feynman-kac formulae with applications to Currently, he is a Senior Researcher with the
non-linear filtering,” in Séminaire de Probabilités XXX4. New York: Laboratory of Computational Engineering, Helsinki
Springer-Verlag, 2000. University of Technology, and a R&D manager with
[37] M. Simandl, J. Královec, and T. Söderström, “Anticipative grid de- Indagon, Ltd. His research interests are in estimation
sign in point-mass approach to nonlinear state estimation,” IEEE Trans. of stochastic systems, Bayesian methods in signal
Autom. Control, vol. 47, pp. 699–702, Apr. 2002. processing, and stochastic control theory.

Anda mungkin juga menyukai