Iterated posterior linearization filters and smoothers with cross-correlated noises

Iterated posterior linearization filters and smoothers with cross-correlated noises

Journal Pre-proof Iterated posterior linearization filters and smoothers with cross-correlated noises Yanhui Wang, Hongbin Zhang, Yang Li PII: DOI: R...

2MB Sizes 0 Downloads 25 Views

Journal Pre-proof Iterated posterior linearization filters and smoothers with cross-correlated noises Yanhui Wang, Hongbin Zhang, Yang Li

PII: DOI: Reference:

S0019-0578(20)30003-3 https://doi.org/10.1016/j.isatra.2020.01.008 ISATRA 3449

To appear in:

ISA Transactions

Received date : 21 May 2018 Revised date : 11 December 2019 Accepted date : 3 January 2020 Please cite this article as: Y. Wang, H. Zhang and Y. Li, Iterated posterior linearization filters and smoothers with cross-correlated noises. ISA Transactions (2020), doi: https://doi.org/10.1016/j.isatra.2020.01.008. This is a PDF file of an article that has undergone enhancements after acceptance, such as the addition of a cover page and metadata, and formatting for readability, but it is not yet the definitive version of record. This version will undergo additional copyediting, typesetting and review before it is published in its final form, but we are providing this version to give early visibility of the article. Please note that, during the production process, errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

© 2020 Published by Elsevier Ltd on behalf of ISA.

*Title page showing Author Details

Journal Pre-proof Title: Iterated

Posterior

Linearization

Filters

and

Smoothers

with

Cross-Correlated Noises

pro of

Author names and affiliations: Yanhui Wang 1

School of Electrical Engineering and Automation, Hefei University of

Technology, Hefei, Anhui, 230009, P.R. China. 2

School of Information and Communication Engineering, University of

P.R. China.

Corresponding author: Hongbin Zhang

lP

[email protected]

re-

Electronic Science and Technology of China, Chengdu, Sichuan, 611731,

urn a

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, P.R. China.

[email protected]

Jo

Li Yang

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, P.R. China. [email protected]

*Highlights (for review)

Journal Pre-proof 1. This paper represents several novel iterated posterior linearization methods to tackle the estimation problem for nonlinear state-space models with cross-correlated noises, which is troublesome to other

pro of

algorithms. 2. The proposed iterated algorithms feature concise derivation, higher estimation accuracy and stronger robustness.

3. The presented iterated approaches can achieve an effective balance between estimation accuracy and computational complexity by the

Jo

urn a

lP

re-

iterated operations.

Journal Pre-proof *Blinded Manuscript - without Author Details Click here to view linked References

pro of

Iterated Posterior Linearization Filters and Smoothers with Cross-Correlated Noises

Abstract

In the article, several concise and efficient iterated posterior linearization filtering and smoothing methodologies are proposed for nonlinear systems with cross-correlated noises. Based on the Gaussian approximation (GA), the presented methods are derived via performing statistical linear regressions (SLRs) of the nonlinear state-space models w.r.t the current posterior distribution in

re-

an iterated way. Various posterior linearization methods can be developed by employing different approximation computation approaches for the Gaussian-weighted integrals encountered in SLRs. These new estimation methods enjoy not only the accuracy and robustness of the GA filter but also the lower computational complexity. Estimation performances of the designed methods are illus-

lP

trated and compared with conventional estimation schemes by two common numerical examples. Keywords: Bayesian filtering, Iterated smoothing, Statistical linear regression, Nonlinear systems, Cross-correlated noises

urn a

1. Introduction

In the recent decades, nonlinear state estimation has been gaining widespread attention in many practical applications [1, 2]. In order to conveniently derive the general solution for the nonlinear estimation, the process and measurement noises, i.e., the system or model noises, are always assumed 5

to be independent. For this case, the Bayesian approach [3] is the most prevalent solution, which can recursively calculate the posterior probability density functions (PDFs). However, due to the intractability of multi-dimensional integrals encountered in the Bayesian approach, the closed-form

Jo

expressions of such PDFs are generally unavailable and numerous Gaussian approximation (GA) filtering and smoothing methods [4–9] are developed to get the approximative numerical solutions. 10

It is worth noting that the GA filters can accurately approximate the posterior PDF only when the posterior PDF is unimodal.

Preprint submitted to Journal of ISA Transactions Templates

December 11, 2019

Journal Pre-proof

However, in many practical engineering applications, the independence assumption of system noises can not be guaranteed and they may be cross-correlated simultaneously or one (sampling) period apart, which will severely weaken the performances of the conventional estimation methods based on the noise independence assumption. Such estimation problem with cross-correlated noises

pro of

15

has drawn considerable attention and become a challenging research topic recently [10–13]. For nonlinear models with synchronous cross-correlated noises, such as the measurement-based output feedback closed-loop control systems, there are three filtering methods including the de-correlating filter [14], the Gaussian approximation (GA) recursive filter (GARF) [15] and the correlated Gaus20

sian approximate filter (CGAF) [16] that can be used to solve this estimation problem. Research proves that the conventional de-correlating filter is equivalent to GARF for linear Gaussian systems [17], and, for nonlinear systems, they are equal to each other in the linear minimum mean

re-

square error (LMMSE) sense [18]. The accuracy of CGAF has proven to be superior to those of its corresponding de-correlating filter and GARF in some applications when the Gaussian-weighted 25

integrals involved in them are calculated by the same Gaussian approximate rules [16]. However, CGAF is sensitive to the correlation parameter. In other words, the advantages of estimation ac-

lP

curacy and robustness can not be guaranteed when the correlation parameter changes only a little, which seriously limits its scope of applications.

Furthermore, the cases that system noises are one-step cross-correlated are also common in 30

many engineering applications. For instance, to discretize a continuous nonlinear system, an addi-

urn a

tional noise contribution will be introduced into the measurement model from the discrete sampling process of the state [19]. Another well-known instance is an airplane flying through the random wind field. The effect of wind in the airplane’s acceleration can be modeled as the process noise, which will be one-step correlated with the measurement noise because any anemometer mounted 35

on the airplane is also corrupted by wind [20]. In order to deal with such estimation problem possessing asynchronous cross-correlated noises, a linear Kalman filter (KF) solution is derived for the linear systems [19, 20], while the systematic solutions to the nonlinear models are rarely studied. Despite the existing Gaussian approximate approaches [21] can be used to solve this kind of

40

Jo

nonlinear estimation problem, such methods will be confronted with the low estimation accuracy and weak robustness, especially for some applications with unimodal Gaussian PDF. Therefore, when confronted with such estimation problem, there is a great demand on exploring more efficient methods to improve the estimation accuracy and robustness.

2

Journal Pre-proof

Besides the standard nonlinear GA filters and smoothers [4–9], there are many iterated estimation ways to obtain the approximative numerical solutions of the Bayesian method, such as iterated 45

extended KF (IEKF) [4, 14], iterated unscented KF (IUKF) [22], iterated posterior linearization

pro of

filter (IPLF) [23], and iterated posterior linearization smoother (IPLS) [24]. The primary distinction between IEKF (IUKF) and EKF (UKF) [4, 5] is the measurement update. The former updates the state to the maximum a posteriori (MAP) estimation obtained by one measurement iteration approximation, whereas the latter calculates the updated state as a linear combination of predic50

tion and innovation, i.e., an approximate conditional mean. Correspondingly, the filtering mean is approximated by a LMMSE-based estimator in the state update of IPLF and the filtering error is approximated as the mean square error (MSE). LMMSE-based filters generally have better filtering accuracy than MAP-based approximations since the estimation performances are usually assessed

55

re-

by square error [23]. Only the measurement functions are relinearized in IPLF while the process models are also relinearised in IPLS [24], which is the crucial distinction between IPLF and IPLS. It has been proved that the iterated approaches are far superior to the corresponding non-iterative methods in many practical applications [25–27].

lP

Enlightened by the advantages of IPLF, an improved estimation performance may be anticipated for nonlinear models possessing cross-correlated noises if statistical linear regressions (SLRs) of 60

the system functions are performed with respect to (w.r.t) the current optimal distribution. Specifically, for the case of simultaneous cross-correlated noises, a de-correlating IPLF can be derived

urn a

directly for nonlinear models since the noise correlation can be eliminated by reconstructing a pseudo process equation in the de-correlating framework and then the conventional nonlinear GA filters can be applied. For the case of one-step cross-correlated noises, some modifications are needed 65

before IPLF is implemented to obtain the nonlinear estimation solution. Firstly, a statistical linear error propagation is derived by implementing SLRs of nonlinear systems and the multidimensional integrals involved in the SLRs are approximated by the GA approaches [5–7]. Secondly, the proposed recursion is executed based on the latest posterior approximation in an iterative way. The motivation comes from the fact that SLRs of nonlinear functions make it easy for the recursion of state estimation with cross-correlated noises to be derived, and the posterior iterated operator

Jo

70

provides a more accurate estimation. Based on these improvements, an effective balance between computational complexity and precision can be expected for these newly proposed approaches. The rest of this paper is outlined below. The general recursive frameworks are derived for

3

Journal Pre-proof

nonlinear models possessing cross-correlated noises in Section 2. In Section 3, a de-correlating IPLF 75

and a modified IPLF are derived for nonlinear models possessing simultaneous and one-step crosscorrelated noises, respectively. To further improve the estimation performance, the corresponding

pro of

smoothing approaches are also proposed. In Section 4, the estimation performances are illustrated and analysed by two common numerical examples. Section 5 draws some conclusions and the further recommendation is presented in Section 6.

80

2. General recursive frameworks for nonlinear models with cross-correlated noises The considered discrete nonlinear systems are modeled by the process (dynamic) and measurement (observation) formulas below

(1)

zk = h(xk ) + vk ,

(2)

re-

xk = f (xk−1 ) + wk−1

where the mappings f (·) and h(·) are nonlinear, xk is an nx -dimensional vector to be estimated, zk

lP

is an nz -dimensional observation, the white noises wk−1 and vk are Gaussian and have covariances Qk−1 and Rk , respectively. The system noises may be cross-correlated simultaneously satisfying E[wk vlT ] = Sk δkl or one-step cross-correlated satisfying E[wk−1 vlT ] = Sk δkl . Here E stands for the expected value operator and δkl denotes the Kronecker delta. The prior PDF is p(x0 ) = ¯ 0 , P0 ), which is uncorrelated with system noises. N (x0 ; x

urn a

The estimation goal is to get the recursive Gaussian approximation (GA) distributions N (xk ; ˆ k|k−1 , Pk|k−1 ) and N (xk ; x ˆ k|k , Pk|k ) to the predicted function p(xk |Zk−1 ) and posterior function x

p(xk |Zk ), respectively. Here, the set Zk = {zi }ki=1 . In other words, the main task is to compute the first two moments of the above two Gaussian distributions as accurate as possible. When system noises are cross-correlated concurrently, the desired state estimation can be transformed into an equivalent estimation with uncorrelated noises by reconstructing a pseudo-process equation [28]. Add a zero vector to the right of formula (1) to get

Jo

85

xk = f (xk−1 ) + wk−1 + Dk−1 (zk−1 − hk−1 (xk−1 ) − vk−1 )

(3)

This results in the modified process and measurement equations xk = (f (xk−1 ) − Dk−1 hk−1 (xk−1 )) + Dk−1 zk−1 + (wk−1 − Dk−1 vk−1 ) 4

(4)

Journal Pre-proof

zk = h(xk ) + vk ,

(5)

which has new process noise of (wk−1 − Dk−1 vk−1 ) with zero mean and the term Dk−1 zk−1 can be

pro of

regarded as an additional deterministic input. The motivation behind these manipulations is that the matrix Dk−1 can be selected to make (wk−1 − Dk−1 vk−1 ) uncorrelated with vk . That is to say   T E (wk−1 − Dk−1 vk−1 )vk−1 = Sk−1 − Dk−1 Rk−1

which is zero if we select

−1 Dk−1 = Sk−1 Rk−1

(6)

(7)

With this choice, the modified process and measurement formulas can be described as (8)

zk = h(xk ) + vk .

(9)

The new process noise autocorrelation is

re-

−1 −1 −1 xk = (f (xk−1 ) − Sk−1 Rk−1 hk−1 (xk−1 )) + Sk−1 Rk−1 zk−1 + (wk−1 − Sk−1 Rk−1 vk−1 )

(10)

lP

  −1 T E (wk−1 − Dk−1 vk−1 )(wk−1 − Dk−1 vk−1 )T = Qk−1 − Sk−1 Rk−1 Sk−1 ,

which is less than the Qk−1 term arising from the original process noise wk−1 .

Enlightened by the KF’s estimation solution to the linear systems possessing one-step crosscorrelated noises, a natural idea is that the state estimation in nonlinear case can be tackled if the

urn a

nonlinear functions can be accurately approximated in linear form. A common approximation of nonlinear function used in the literature is shown below f (xk ) ≈ Fk xk + ak + ek

(11)

h(xk ) ≈ Hk xk + bk + gk ,

(12)

where the linearization parameters (Fk , Hk , and so forth) have appropriate dimensions, ek and gk are Gaussian white noises and their covariances are Λk and Ωk , respectively, which can be regarded as the terms for assessing the approximation errors.

Jo

90

In the enabling linear approximation of nonlinear function with associated PDF, SLR method provides a linearization approximation (affine) function with minimum mean square error (MMSE). SLR has the ability to obtain the affine function with high linearization approximate precision, which is described as below.

5

Journal Pre-proof

¯ P ), SLR of f (x) Definition 1. For nonlinear mapping f (x) with Gaussian PDF p(x) = N (x; x, w.r.t p(x) can be summarized as follows [29]

(14)

Λ = Φ − F P F T,

(15)

Φ= 95

Z

Z

f (x)p(x)dx

(16)

¯ ¯ T p(x)dx (x − x)(f (x) − y)

(17)

¯ (x) − y) ¯ T p(x)dx. (f (x) − y)(f

(18)

re-

Ψ=

Z

(13)

¯ a = y¯ − H x

where y¯ =

pro of

F = ΨT P −1

For nonlinear mapping f (x), a best affine approximation f (x) ≈ F x + a can be achieved with MMSE and Λ is its mean square error (MSE) matrix [29].

In most practice nonlinear cases, computing the required moments of SLR (16)–(18) directly

lP

is intractable and the multidimensional integrals can be approximately calculated by the sigmapoint approaches. One frequently used approximation technique is shown as Algorithm 1, whose sigma-points and weights are selected according to a suitable GA method [5–7]. Algorithm 1 SLR of nonlinear function using Gaussian approximation approach ¯ P. Input: Nonlinear function f (x) and a Gaussian PDF p(x) possessing the first two moments x,

urn a

Output: The returned affine approximation parameters (F , a, Λ). ¯ and P , and select the corresponding weights w1 , . . . , wm . 1: Generate m sigma-points χ1 , . . . , χm according to x 2: Estimate the propagated sigma-points Yi = f (χi )

i = 1, . . . , m.

3: Approximate the moments (16)–(18) as

y¯ ≈

Jo

Ψ≈ Φ≈

m X i=1 m X

i=1 m X i=1

wi Yi ¯ ¯ T wi (χi − x)(Y i − y) ¯ ¯ T wi (Yi − y)(Y i − y)

4: Calculate (F , a, Λ) from equations (13)–(15). 100

¯ P ) to stand for SLR of f (x) w.r.t For notational clarity, we use (F , a, Λ) = SLR(f (x), x, ¯ P ) and (F , a, Λ) are the returned approximation parameters. Gaussian PDF N (x; x, 6

Journal Pre-proof

ˆ k|k by updating the previous estimaTo proceed on, it is desired to get the current estimation x ˆ k−1|k−1 based on all currently available measurements. With the aid of linear approximation, tion x 105

the general recursive estimation framework can be derived directly for nonlinear models possessing

pro of

one-step cross-correlated noises, which is summarized as Theorem 1.

ˆ k|k , Pk|k ) of p(xk |Zk ) is obtained by the formulas below Theorem 1. The GA distribution N (x; x ˆ k|k = x ˆ k|k−1 + Kk (zk − Hk x ˆ k|k−1 − bk ) x

 Pk|k = Pk|k−1 − Kk Hk Pk|k−1 HkT + Ωk + Rk + Hk Sk + SkT HkT KkT −1 Kk = (Pk|k−1 HkT + Sk ) Hk Pk|k−1 HkT + Ωk + Rk + Hk Sk + SkT HkT

re-

where Kk is gain and

ˆ k|k−1 = Fk−1 x ˆ k−1|k−1 + ak−1 x T Pk|k−1 = Fk−1 Pk−1|k−1 Fk−1 + Λk−1 + Qk−1

ˆ k|k−1 , Pk|k−1 ) (Hk , bk , Ωk ) = SLR(h(x), x

lP

ˆ k−1|k−1 , Pk−1|k−1 ). (Fk−1 , ak−1 , Λk−1 ) = SLR(f (x), x Proof. The analogous theorem has been proved in the relevant literature [14, 19]. Similar to Theorem 1, for nonlinear models possessing simultaneous cross-correlated noises, the analogous recursive framework can be obtained immediately by applying the posterior linearization

urn a

filter (PLF) [23] to the equations (8) and (9), which is depicted as Theorem 2. ˆ k|k , Pk|k ) by the Theorem 2. The GA of p(xk |Zk ) is expressed as a normal distribution N (x; x following formulae

ˆ k|k = x ˆ k|k−1 + Kk (zk − Hk x ˆ k|k−1 − bk ) x

 Pk|k = Pk|k−1 − Kk Hk Pk|k−1 HkT + Ωk + Rk KkT −1 Kk = Pk|k−1 HkT Hk Pk|k−1 HkT + Ωk + Rk

Jo

110

where Kk is gain and

−1 Dk−1 = Sk−1 Rk−1 ∗ Fk−1 = Fk−1 − Dk−1 Hk−1

7

Journal Pre-proof

∗ ˆ k|k−1 = Fk−1 ˆ k−1|k−1 + Dk−1 (zk−1 − bk−1 ) + ak−1 x x ∗ ∗ T Pk|k−1 = Fk−1 Pk−1|k−1 (Fk−1 )T + Λk−1 + Qk−1 + Dk−1 (Ωk−1 − Rk−1 )Dk−1

pro of

ˆ k|k−1 , Pk|k−1 ) (Hk , bk , Ωk ) = SLR(h(x), x ˆ k−1|k−1 , Pk−1|k−1 ). (Fk−1 , ak−1 , Λk−1 ) = SLR(f (x), x

If the affine functions can accurately approximate nonlinear system functions, i.e., the terms Λ = 0 and Ω = 0, an alternative filtering formulation of Theorem 2, which produces a priori mean and covariance, can be expressed as follows after some tedious algebraic manipulations

re-

ˆ k+1|k = Fk x ˆ k|k−1 + Kk (zk − Hk x ˆ k|k−1 − bk ) + ak x −1 T Kk + Qk−1 Pk+1|k = Fk Pk|k−1 FkT − Kk Hk Pk|k−1 HkT + Rk  −1 Kk = (Fk Pk|k−1 HkT + Sk ) Hk Pk|k−1 HkT + Rk .

(19) (20) (21)

Here Kk is called the modified Kalman gain. Except for formula (21), equations (19) and (20) are identical to their counterparts of the filter with independent system noises, which can be obtained

lP

directly from equations (19)–(21) by setting the cross-correlation term Sk = 0. The only effect of the simultaneous correlation of system noises is to modify the Kalman gain. From equation (20), 115

it is clear that the error covariance is decreased owing to more useful information provided by the cross-correlation term Sk .

urn a

3. Development of the filters and smoothers with correlated noises 3.1. The de-correlating iterated estimation methods with simultaneous cross-correlated noises Inspired by the development of de-correlating estimation framework as well as the superiority of 120

IPLF, a de-correlating IPLF method can be derived for nonlinear models possessing simultaneous ˆ 0|0 = x0 and P0|0 = P0 . The cross-correlated noises. Assume the initial estimation moments are x de-correlating IPLF algorithm is summarized as Algorithm 2.

Jo

To take full advantage of the measurement information, SLR of observation function is executed w.r.t current posterior estimation in each iteration (except the first iteration) update of Algorithm 2. 125

In the de-correlating IPLF, the best available approximation of posterior distribution is taken into account to perform iterated SLRs of nonlinear functions.

8

Journal Pre-proof

Algorithm 2 The de-correlating iterated posterior linearization filter ˆ 0|0 , P0|0 and the number of iterations J. Input: Initial moments x ˆ k|k , Pk|k (k = 0, . . . , N ). Output: Posterior moment estimations x 1: for k = 1 → N do

1 =P ˆ k|k ˆ k|k , Pk|k by Theorem 2 and set Pk|k ˆ 1k|k = x Compute x k|k , x

3:

for j = 2 → J do

4:

j−1 ˆ j−1 (Hkj , bjk , Ωjk ) = SLR(h(x), x , Pk|k ) k|k

5:

ˆ jk|k x

6:

j−1 Kkj = Pk|k (Hkj )T j j−1 Pk|k = Pk|k − Kkj

7:

=

ˆ j−1 x k|k

+

Kkj (zk h



ˆ j−1 Hkj x k|k



pro of

2:

. First iteration

. Using Algorithm 1

bjk )

i−1 j−1 Hkj Pk|k (Hkj )T + Ωjk + Rk h i j−1 Hkj Pk|k (Hkj )T + Ωjk + Rk (Kkj )T

8:

end for

9:

J . ˆ k|k = x ˆJ Set x , Pk|k = Pk|k k|k

re-

10: end for

Based on a reconstructed pseudo process equation and Rauch-Tung-Striebel (RTS) smoother [30], the de-correlating RTS iterated posterior linearization smoother (RTS-IPLS) algorithm can be obtained directly for nonlinear models possessing simultaneous cross-correlated noises, which is sum-

lP

marized as Algorithm 3.

Algorithm 3 The de-correlating RTS iterated posterior linearization smoother ˆ 0|0 , P0|0 and the number of iterations J. Input: Initial moments x Output: Posterior moment estimations uk|k , Wk|k (k = 0, . . . , N ). ˆ k|k , Pk|k by Algorithm 2. 1: Compute the first two moments x ˆ N |N , WN |N = PN |N . 2: uN |N = x 4: 5: 6:

−1 Ck = Pk|k (Fk∗ )T Pk+1|k

ˆ k|k + Ck (uk+1|k+1 − x ˆ k+1|k ) uk|k = x

Wk|k = Pk|k + Ck (Wk+1|k+1 − Pk+1|k )CkT

7: end for 130

. Smoothing

urn a

3: for k = N − 1 to 0 do

. Filtering

3.2. The proposed iterated filter and smoother with one-step cross-correlated noises

Jo

On account of Theorem 1, for nonlinear models possessing one-step cross-correlated noises, a modified IPLF method can be derived in a similar way to Algorithm 3, which is summarized in Algorithm 4. Considering that the effect of cross-correlated noises has been partially incorporated 135

into the step 2 of Algorithm 4 and the subsequent iterations primarily rely on the measurements, the Kalman gain is modified as step 6 of Algorithm 4 in the following iterations. 9

Journal Pre-proof

Algorithm 4 The modified iterated posterior linearization filter ˆ 0|0 , P0|0 and the number of iterations J. Input: Initial moments x ˆ k|k , Pk|k (k = 0, . . . , N ). Output: Posterior moment estimations x 1: for k = 0 → N do

1 =P ˆ k|k ˆ k|k , Pk|k by Theorem 1 and set Pk|k ˆ 1k|k = x Compute x k|k , x

3:

for j = 2 → J do

4:

j−1 ˆ j−1 (Hkj , bjk , Ωjk ) = SLR(h(x), x , Pk|k ) k|k

5:

ˆ jk|k x

6:

j−1 Kkj = Pk|k (Hkj )T j j−1 Pk|k = Pk|k − Kkj

7:

=

ˆ j−1 x k|k

+

Kkj (zk h



ˆ j−1 Hkj x k|k



pro of

2:

. First iteration

. Using Algorithm 1

bjk )

i−1 j−1 Hkj Pk|k (Hkj )T + Ωjk + Rk + Hkj Sk + SkT (Hkj )T h i j−1 Hkj Pk|k (Hkj )T + Ωjk + Rk + Hkj Sk + SkT (Hkj )T (Kkj )T

8:

end for

9:

J . ˆ k|k = x ˆJ Set x , Pk|k = Pk|k k|k

re-

10: end for

Unlike the de-correlating RTS-IPLS, the RTS smoother no longer holds for nonlinear models possessing one-step cross-correlated noises in the sense of mathematics. Fortunately, on the basis of SLR approximation of nonlinear systems, the desired smoother can be interpreted as a function of two linear filters [31], which results in a so-called forward-backward posterior linearization smoother

lP

(FB-PLS). One of these two linear filters is the forward filter that works forward over measurements to the point of interest, which has been shown as Theorem 1. The other one referred to as the backward information filter runs backward over measurements to the same point, which is described as Theorem 3.

urn a

Theorem 3. The GA of posterior distribution p(xb,k |Zk ) using the information vector yb,k|k and matrix Yb,k|k has the following form

T Yb,k|k = Yb,k|k−1 +(Hb,k +Yb,k|k−1 Sk ) Rk +Ωb,k −SkT Yb,k|k−1 Sk T Lb,k = (Hb,k + Yb,k|k Sk )(Rk + Ωb,k + Hb,k Sk )−1  yb,k|k = I + Lb,k SkT yb,k|k−1 + Lb,k (zk − bb,k )  −1 Mb,k = Yb,k+1|k+1 Yb,k+1|k+1 + (Qk + Λb,k )−1

T yb,k|k−1 = Fb,k (I − Mb,k ) yb,k+1|k+1 − Yb,k+1|k+1 ab,k

Jo

140

T Yb,k|k−1 = Fb,k (I − Mb,k )Yb,k+1|k+1 Fb,k

ˆ b,k+1|k+1 , Pb,k+1|k+1 ) (Fb,k , ab,k , Λb,k ) = SLR(f (x), x ˆ b,k|k−1 , Pb,k|k−1 ). (Hb,k , bb,k , Ωb,k ) = SLR(h(x), x 10



−1

(SkT Yb,k|k−1 +Hb,k )

Journal Pre-proof

To clarify the above notations, we first define the following covariance expressions

where ˜ b,k|k−1 , x ˆ b,k|k−1 − xk , x 145

h i ˜ b,k|k x ˜T Pb,k|k , E x b,k|k

pro of

h i ˜ b,k|k−1 x ˜T Pb,k|k−1 , E x b,k|k−1 ,

˜ b,k|k , x ˆ b,k|k − xk . x

(22)

(23)

Just before the measurement at time tk , the backward state xb,k is estimated by the backwardˆ b,k|k−1 and x ˆ b,k|k are the backward filtering expectation and prediction running filter. Hence, x mean, respectively. Due to the measurement going backward in time, the measurement update and prediction roles of backward filter are reversed from forward filter. The information form of −1 backward filter is the backward information filter and they are related by Yb,k|k = Pb,k|k , Yb,k|k−1 =

re-

−1 ˆ b,k|k−1 . ˆ b,k|k , and yb,k|k−1 = Yb,k|k−1 x Pb,k|k−1 , yb,k|k = Yb,k|k x

Proof. We only prove the propagations of Yb,k|k and yb,k|k . The proofs of the remaining equations are easily derived and omitted for convenience. Based on Theorem 1, the covariance Pb,k|k can also be denoted as

lP

 −1 T T T Pb,k|k = Pb,k|k−1 −(Pb,k|k−1 Hb,k +Sk ) Hb,k Pb,k|k−1 Hb,k +Ωb,k +Rk +Hb,k Sk +SkT Hb,k (Hb,k Pb,k|k−1 +SkT )

By utilizing the matrix inversion lemma, an alternative form of Pk|k is written as  −1  −1 −1 −1 −1 −1 T Pb,k|k = Pb,k|k−1 + (Hb,k + Pb,k|k−1 Sk ) Rk + Ωb,k − SkT Pb,k|k−1 Sk (SkT Pb,k|k−1 + Hb,k )

urn a

Inversing both sides of the above equation and substituting the information matrix into this resulting matrix inversion expression yields

T Yb,k|k = Yb,k|k−1 + (Hb,k + Yb,k|k−1 Sk ) Rk + Ωb,k − SkT Yb,k|k−1 Sk

−1

(SkT Yb,k|k−1 + Hb,k )

To derive the expression of yb,k|k , the backward Kalman gain Kb,k can be obtained directly based on Theorem 1, which is described as

  T T T Kb,k Rk + Ωb,k + Hb,k Sk + Hb,k Pb,k|k−1 Hb,k + SkT Hb,k = (Pb,k|k−1 Hb,k + Sk )

Jo

150

Collecting terms now gives

 T  + Sk Kb,k (Rk + Ωb,k + Hb,k Sk ) = (I − Kb,k Hb,k )Pb,k|k−1 − Kb,k SkT Hb,k T = Pb,k|k Hb,k + Sk

11

Journal Pre-proof

The above expression can be reorganized as T Kb,k = (Pb,k|k Hb,k + Sk )(Rk + Ωb,k + Hb,k Sk )−1

pro of

Then one can write the backward measurement update equation as ˆ b,k|k = x ˆ b,k|k−1 + Kb,k (zk − Hb,k x ˆ b,k|k−1 − bb,k ) x

−1 ˆ b,k|k−1 + Kb,k (zk − bb,k ) = (Pb,k|k + Kb,k SkT )Pb,k|k−1 x

−1 −1 ˆ b,k|k−1 + Kb,k (zk − bb,k ) = Pb,k|k (I + Pb,k|k Kb,k SkT )Pb,k|k−1 x −1 Left multiplying Pb,k|k on both sides of the above formula and substituting the information vector

and matrix into this resulting matrix inversion expression lead to the desired form

re-

 yb,k|k = I + Lb,k SkT yb,k|k−1 + Lk (zk − bb,k )

T with Lb,k = (Hb,k + Yb,k|k Sk )(Rk + Ωb,k + Hb,k Sk )−1 .

lP

Forward filter

ˆ k|k−1 x ˆ k|k x yb,k|k yb,k|k−1 k

N

Backward information filter

urn a

0

Figure 1: Forward filter and backward information filter.

The forward filter and backward information filter are illustrated as Figure 1. The task is to ˆ k|k and yb,k|k−1 . Since the forward filter and seek a smoothing estimation that is a combination of x backward information filter are uncorrelated, based on the Millman’s theorem [32], the smoothing

Jo

estimation µk|k can be obtained directly and expressed as −1 −1 ˆ k|k + Pb,k|k−1 ˆ b,k|k−1 ) µk|k = Wk|k (Pk|k x x

(24)

with the smoothing covariance −1 −1 Wk|k = (Pk|k + Pb,k|k−1 )−1

12

(25)

Journal Pre-proof

For scalar systems, equation (25) reduces down to Wk|k =

Pk|k Pb,k|k−1 Pk|k + Pb,k|k−1

(26)

pro of

Equation (26) clearly shows that Wk|k ≤ Pk|k and Wk|k ≤ Pb,k|k−1 , which indicates that, in most cases, the smoothing accuracy is always higher than either the forward filter or the backward filter. This analysis and conclusion can easily be extended to the multidimensional nonlinear systems. Since the smoother and the forward filter must have the same estimation at terminal time tk = tN , ˆ N |N and WN |N = PN |N . According to equation (25) then, this obviously requires that µN |N = x

−1 Pb,N |N −1 ≡ Yb,N |N −1 = 0. However, for the following backward measurement update equation (27),

ˆ b,N |N −1 is yet unknown the terminal state boundary x

(27)

re-

ˆ b,k|k = x ˆ b,k|k−1 + Kb,k (zk − Hb,k x ˆ b,k|k−1 − bb,k ) x

To solve this difficulty, the information form of state update is considered, which is given in Theˆ b,k|k−1 we have orem 3. Since Yb,N |N −1 = 0, then from the relation equation yb,k|k−1 = Yb,k|k−1 x ˆ b,N |N −1 . This is why we propagate yb,k|k−1 instead yb,N |N −1 = 0, which is valid for any value of x

lP

ˆ b,k|k−1 in backward information filter since we do not know the boundary condition x ˆ b,N |N −1 . of x After some algebraic manipulations, the equations (24) and (25) are respectively rewritten as ˆ k|k + Wk|k yb,k|k−1 uk|k = (I − Gk )x

(29)

urn a

Wk|k = (I − Gk )Pk|k

(28)

with the smoothing gain Gk = Pk|k Yb,k|k−1 (I + Pk|k Yb,k|k−1 )−1 . Based on the aforementioned discussions, the forward-backward iterated posterior linearization smoother (FB-IPLS) algorithm can be derived as Algorithm 5. The correlation parameters Sk are not taken into account in the iterated 155

operations of Algorithm 5 since the effect of the noise cross-correlation has been incorporated into the anterior smoothing manipulations.

It is noteworthy that, in the iterated operations of Algorithms 2 and 3, the SLRs of nonlin-

Jo

ear functions are performed w.r.t the latest eatimation which only considers previous and current measurements. In Algorithm 5, the future measurements are also taken into account by the s160

moothing operations to linearize nonlinear systems and the estimated performances are able to be further improved in theory since the future measurements can also provide useful information. It also should be noted that each iteration (except the first iteration) of Algorithm 5 only relinearizes 13

Journal Pre-proof

Algorithm 5 The forward-backward iterated posterior linearization smoother ˆ 0|0 , P0|0 and the number of iterations J. Input: Initial moments x Output: Posterior moment estimations uk|k , Wk|k (k = 0, . . . , N ). ˆ k|k , Pk|k using Theorem 1. 1: Compute x

pro of

. Forward filtering

ˆ N |N , WN |N = PN |N , yb,N |N −1 = 0, Yb,N |N −1 = 0. 2: Smoothing initialisation: uN |N = x 3: for k = N − 1 to 0 do 4:

Calculate the information vector yb,k|k and matrix Yb,k|k by Theorem 3. )−1

5:

Compute the smoothing gain Gk = Pk|k Yb,k|k−1 (I + Pk|k Yb,k|k−1

6:

Evaluate the estimation covariance Wk|k = (I − Gk )Pk|k

7:

ˆ k|k + Wk|k yb,k|k−1 Calculate the estimation state uk|k = (I − Gk )x

8: end for

1 =W 9: Iteration initialisation: u1k|k = uk|k , Wk|k k|k (k = 0, . . . , N ).

10: for j = 2 → J do 11:

for k = 0 → N do

j−1 (Hkj , bjk , Ωjk ) = SLR(h(x), uj−1 , Wk|k ) k|k

13:

− bjk ) + Kkj (zk − Hkj uj−1 ujk|k = uj−1 k|k k|k h i−1 j−1 j−1 Kkj = Wk|k (Hkj )T Hkj Wk|k (Hkj )T + Ωjk + Rk i h j−1 j−1 j (Hkj )T + Ωjk + Rk (Kkj )T − Kkj Hkj Wk|k = Wk|k Wk|k

14: 15: 16:

re-

12:

end for

. Smoothing

. Iteration

. Using Algorithm 1

lP

17: end for

. Backward filtering

J (k = 0, . . . , N ). 18: Set uk|k = uJ , Wk|k = Wk|k k|k

the measurement functions rather than both the dynamic and measurement functions [24]. The

165

urn a

reasons for this are that the subsequent iterations depend mainly on the measurements and the computational burden can be reduced significantly by this simplified linearizing operation. Notice that the backward SLR parameters (Fb,k , Hb,k and so forth) need to be calculated in Algorithm 5, they are generally approximated and substituted by the corresponding forward SLR parameters to save computing overhead when the approximation error is small. Although RTS smoother no longer holds in the strict mathematical sense when the system 170

noises are one-step cross-correlated, compared with any loss in estimation quality, the savings

Jo

in computational load make this smoothing approach attractive in estimating the state of many nonlinear systems, which results in a suboptimal but simplified smoother. When the cross-correlated parameters Sk are small, the suboptimal RTS iterated posterior linearization smoother (RTSIPLS) can be obtained by substituting the RTS smoothing manipulations into Algorithm 5, which 175

is described as Algorithm 6. Compared with Algorithm 5, the calculation error can be further 14

Journal Pre-proof

decreased in Algorithm 6 without calculating the backward information filter. Algorithm 6 The Rauch-Tung-Striebel iterated posterior linearization smoother ˆ 0|0 , P0|0 and the number of iterations J. Input: Initial moments x ˆ k|k , Pk|k by Theorem 1. 1: Compute posterior moments x

pro of

Output: Posterior moment estimations uk|k , Wk|k (k = 0, . . . , N ). ˆ N |N , WN |N = PN |N . 2: Smoothing initialisation: uN |N = x 3: for k = N − 1 to 0 do 4:

−1 Ck = Pk|k FkT Pk+1|k

5:

ˆ k|k + Ck (uk+1|k+1 − x ˆ k+1|k ) uk|k = x

6:

Wk|k = Pk|k + Ck (Wk+1|k+1 − Pk+1|k )CkT

7: end for

8: The iteration steps are the same as those of Algorithm 5 counterparts.

. Filtering . Smoothing

. Iteration

re-

The iteration processes of Algorithms 2–6 will be stopped only when certain termination conditions are satisfied. Termination conditions may be different in various situations, and the number of iterations is a common criterion for terminating the iteration process. Since a good estimation 180

performance can generally be obtained with an acceptable computational cost in many practical

lP

applications only after 2–3 iterations, the value of iterations J is typically set to value 2–3. Another ˆ jk|k − x ˆ j−1 criterion in common use for terminating the iteration is that the inequality kx k|k k ≤ Vth can be met. The predefined threshold Vth plays an important role in successfully use of the iterated method, but it is not easy to select an appropriate Vth . Furthermore, a modified inequality for 185

terminating the iteration process is given in [22]. Although after some iterations, the modified

urn a

inequality does not hold automatically, the number of iterations is usually a considerable value. What’s worse, the modified inequality may not work well in some cases because it is sensitive to numerical errors.

Remark 1. Algorithm 2 is similar to the existing filters with cross-correlated noises mentioned 190

in the introduction and it can be reduced to them under certain conditions [23]. More precisely, if Λ = 0, Ω = 0 and the linearization parameters F , H, a and b are obtained based on Taylor

Jo

series expansions at MAP estimation, Algorithm 2 is reduced to IEKF [25] for nonlinear models possessing simultaneous cross-correlated noises. If J = 1 and the SLR parameters F , H, a, b, Λ, and Ω are selected w.r.t the prior, Algorithm 2 is reduced to prior linearization filter (PrLF) [23]. If 195

the sigma-points that are drawn from the prior estimation are applied to approximating the moments of PrLF , Algorithm 2 becomes the common derivative-free KF such as UKF [5] or CKF [6]. 15

Journal Pre-proof

3.3. Computational complexity analysis The computation overheads of Algorithms 2–6 and their corresponding comparison approaches are analyzed and compared in this subsection. The de-correlating algorithms IPLF and RTSIPLS are compared with the existing method CGAF [16] for nonlinear models possessing simul-

pro of

200

taneous cross-correlated noises. For nonlinear models possessing one-step cross-correlated noises, the modified IPLF, FB-IPLS, and RTS-IPLS are compared with the GA filtering and smoothing methods [21], which are called as the Gaussian approximate standard filter (GASF) and Gaussian approximate standard smoother (GASS) respectively in this paper. 205

For a fair comparison, the Gaussian-weighted integrals encountered in all algorithms are computed by the cubature rule [6]. To assess the computational complexity, the floating-point operations (FLOPs) required by all the considered algorithms are analyzed in detail based on the technical

re-

report [33]. In the evaluation process, the FLOPs of the arithmetic operations such as the multiplication, the matrix inversion, the computations of system functions, and the Cholesky decomposition 210

are needed to be calculated. It is noteworthy that some steps of the proposed Algorithms 2–6 are not executed since they are not required or they have been implemented in the previous operations.

lP

For example, the SLR equations (14) and (15) are only implemented in Algorithm 5 (FB-IPLS), which result the substituted SLR parameters for the backward information filter. Furthermore, the computations of process and measurement functions are not taken into account since they depend on the different functions. The FLOPs of all considered algorithms are summarized in Table 1. FLOPs

urn a

Algorithms

1. The de-correlating IPLF

19 3 33 2 1 3 2 2 2 3 nx + 2nz + 6nx nz + 8nx nz + 2 nx + 4nz + 11nx nz + 6 nx + 3nz +

5 2 2 (J −1)( 73 n3x +n3z +6n2x nz +8nx n2z + 21 2 nx +nz +11nx nz − 6 nx +nz )

2. The de-correlating RTS-IPLS

47 2 13 52 3 2 2 3 2 3 nx +2nz +6nx nz +8nx nz + 2 nx +4nz +11nx nz + 6 nx +3nz +

5 2 2 (J −1)( 73 n3x +n3z +6n2x nz +8nx n2z + 21 2 nx +nz +11nx nz − 6 nx +nz )

3. CGAF [16]

4n4x + 31 6 nx

4. IPLF

Jo

215

5. FB-IPLS

61 3 3 nx

+ 3n3z + 6n2x nz + 10nx n2z +

83 2 2 nx

+ 7n2z + 12nx nz +

+ 6nz − 1

43 3 7 3 2 2 45 2 2 3 nx +nz +10nx nz +12nx nz + 2 nx +nz +10nx nz + 6 nx +nz +(J −

15 2 7 3 3 2 2 2 1)( 10 3 nx +nz +10nx nz +12nx nz + 2 nx +nz +10nx nz + 6 nx +nz )

91 3 25 3 2 2 53 2 2 3 nx +3nz +24nx nz +20nx nz + 2 nx +5nz +10nx nz + 6 nx +4nz + 7 3 3 2 2 15 2 2 (J−1)( 10 3 nx +nz +10nx nz +8nx nz + 2 nx +nz +10nx nz + 6 nx +nz )

16

Journal Pre-proof

76 3 59 2 19 3 2 2 2 3 nx + nz + 10nx nz + 12nx nz + 2 nx + nz + 10nx nz + 6 nx + nz +

6. RTS-IPLS

7 3 3 2 2 15 2 2 (J−1)( 10 3 nx +nz +10nx nz +8nx nz + 2 nx +nz +10nx nz + 6 nx +nz ) 71 2 2 nx

+ n2z + 23nx nz − 21 nx + nz

29n3x + n3z + 10n2x nz + 12nx n2z +

8. GASS [21]

15 2 2 56n3x +2n3z +47n2x nz +23nx n2z + 105 2 nx +4nz +29nx nz + 2 nx +4nz

pro of

7. GASF [21]

Table 1: The FLOPs of all considered algorithms

Remarkably, in most cases, state dimension is not less than measurement dimension, i.e., nx ≥ nz (nz ≥ 1). It is obvious that the number of FLOPs of the filter is smaller than that of its corresponding smoother for any values of nx (nx ≥ 1) and nz since more computing operations are needed in the smoother. Moreover, the number of FLOPs of the proposed algorithm is smaller than that of its compared approach for most of values of nx and nz when J = 1.

re-

220

In order to simplify the discussion, only the de-correlating IPLF and its corresponding compared approach CGAF are considered. The other proposed algorithms and their corresponding compared approaches can also be analysed in the same way and the analogous conclusions can also be obtained.

lP

Ignoring the first-order terms which have little influence on the FLOPs, the FLOPs of the decorrelating IPLF and its compared algorithm CGAF are simplified as follows 43 3 45 n + n3z + 10n2x nz + 12nx n2z + n2x + n2z + 10nx nz 3 x 2 10 3 15 3 2 + (J − 1)( nx + nz + 10nx nz + 12nx n2z + n2x + n2z + 10nx nz ) 3 2 71 2 3 3 2 2 2 CGAF : 29nx + nz + 10nx nz + 12nx nz + nx + nz + 23nx nz 2

urn a

The de−correlating IPLF :

For the case of J ≥ 2, the comparison is discussed by setting J = 2 for convenience. In order to facilitate analysis, we assume that the state has the same dimension as the measurement, i.e., nx = nz (the simulation example belongs to this case). Collecting terms, the FLOPs of IPLF and CGAF are equal to 225

191 3 3 nx

2 + 52n2x and 52n3x + 119 2 nx , respectively. The ratio of FLOPs of IPLF and

FLOPs of CGAF is about 1.2 when the state dimension nx is large, which is acceptable for many

Jo

practical applications. If nx = 1, the ratio of FLOPs of IPLF and FLOPs of CGAF reduces to about 1, which means that the computational cost of IPLF is approximate to that of CGAF. Then we assume the state dimension is twice as large as that of the measurement, i.e., nz = 12 nx . Collecting terms, the FLOPs of IPLF and CGAF are equal to 230

407 3 12 nx

+

81 2 2 nx

and

297 3 8 nx

+

189 2 4 nx ,

respectively. It is obvious that the FLOPs of IPLF is smaller than that of CGAF for any values of 17

Journal Pre-proof

nx (nx ≥ 2) and nz = 12 nx (nz ≥ 1). The ratio of FLOPs of IPLF and FLOPs of CGAF is about 0.91 when the state dimension nx is large, which means that the computational cost of IPLF is less than that of CGAF. When the number of iterations J ≥ 3, the analogous analyses can also be 235

the subsequent corresponding simulation results.

4. Numerical simulation

pro of

made in the same way. It is remarkable that the aforementioned conclusions are consistent with

The efficiencies of the proposed Algorithms 2–6 are shown for two numerical examples in this section. The de-correlating IPLF and RTS-IPLS are compared with the existing method CGAF [16] for nonlinear models possessing simultaneous cross-correlated noises. Furthermore, when the number of iterations J = 1, the de-correlating IPLF and RTS-IPLS reduce to the de-correlating PLF

re-

240

and RTS-PLS, respectively, which are also taken into account for comparison. Notice that GARF [15] is omitted for convenience since the performance of CGAF is superior to that of GARF [16]. For nonlinear models possessing one-step cross-correlated noises, the modified IPLF, FB-IPLS, and

4.1. A model of trigonometric and exponential functions with simultaneous cross-correlated noises In this subsection, a numerical simulation example of trigonometric and exponential functions frequently encountered in tracking problems [5, 14] is given, whose system model is described as

urn a

follows [15, 16]

  1             xk = x2,k  =  x1,k−1 + e−0.05x3,k−1 + 10  + 1 wk−1       1 0.2x1,k−1 (x2,k−1 + x3,k−1 ) x3,k 

x1,k





3 sin2 (5x2,k−1 )

zk = 2x1,k + cos(x2,k x3,k ) + vk ,



(30)

(31)

where the Gaussian white noises wk and vk are correlated simultaneously with covariance Sk and their variances are Qk = 1 and Rk = 4, respectively. The initial state x0 = [−0.7 1 1]T , and prior

Jo

245

lP

RTS-IPLS algorithms are compared with the algorithms GASF and GASS [21].

ˆ 0|0 = [0 0 0]T and P0|0 = I3 , where I3 represents an identity moment estimations are selected as x matrix of dimension 3.

18

Journal Pre-proof

6

1.8 CGAF

5

De-Correlating RTS-IPLS

1.4 1.2

1 0

4

CGAF De-Correlating PLF De-Correlating IPLF De-Correlating RTS-PLS De-Correlating RTS-IPLS

pro of

RMSE of x

RMSE of x

3

De-Correlating RTS-PLS

De-Correlating IPLF

2

1.6

De-Correlating PLF

3 2

20

40

60

80

Time step k

100

(a)

1 0

20

40

60

80

Time step k

100

(b)

Figure 2: Performances of the considered algorithms for time step k with Sk = 0.5 and J = 2. 1.6

5.7

De-Correlating IPLF

1.3

1.1 1

2

3

Number of iterations J

(a)

1600 1400 1200 1000 800 600 1

CGAF

De-Correlating PLF

De-Correlating RTS-PLS

4

De-Correlating RTS-PLS

5.1

De-Correlating IPLF

De-Correlating RTS-IPLS

4.8

2

De-Correlating RTS-IPLS

3

Number of iterations J

4

4.2 1

5

De-Correlating IPLF

urn a

Number of Divergences

1800

De-Correlating PLF

4.5

lP

1.2

3

De-Correlating RTS-IPLS

ARMSE of x

De-Correlating RTS-PLS

1.4

2000

CGAF

5.4

5

(c)

2

3

Number of iterations J

4

5

(b) Average Computational Time (ms)

ARMSE of x

2

1.5

De-Correlating PLF

re-

CGAF

60 CGAF 50 40

De-Correlating PLF

De-Correlating IPLF

De-Correlating RTS-PLS De-Correlating RTS-IPLS

30 20 10 1

2

3

Number of iterations J

4

5

(d)

Jo

Figure 3: Performances of the considered algorithms for different iterations J with Sk = 0.5. (a) ARMSE in x2 . (b) ARMSE in x3 . (c) Number of Divergences. (d) Average computational time.

19

Journal Pre-proof

1.3

De-Correlating RTS-PLS De-Correlating RTS-IPLS

1 0.9 0.1

0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

0.8

0.9

1

(a)

1.2

1 0.1

Number of Divergences

De-Correlating PLF De-Correlating IPLF

0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

5.2

0.8

0.9

1

De-Correlating RTS-PLS De-Correlating RTS-IPLS 0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

(c)

0.8

0.9

1

De-Correlating PLF De-Correlating IPLF

1600 1300

De-Correlating RTS-PLS

re-

4.8

CGAF

1900

1000

700

400 0.1

lP

ARMSE of x3

1.4

2200

CGAF

4 0.1

De-Correlating IPLF De-Correlating RTS-IPLS

(b)

5.6

4.4

De-Correlating PLF

De-Correlating RTS-PLS

1.6

pro of

1.1

CGAF

De-Correlating IPLF

ARMSE of x2

ARMSE of x1

1.2

1.8

De-Correlating PLF

CGAF

De-Correlating RTS-IPLS

0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

0.8

0.9

1

(d)

Figure 4: Performances of the considered algorithms for different correlation parameters Sk with J = 2. (a) ARMSE in x1 . (b) ARMSE in x2 . (c) ARMSE in x3 . (d) Number of Divergences.

urn a

The root MSE (RMSE) and accumulative RMSE (ARMSE) are selected as performance metrics and can be expressed as follows based on the state component xi,k (i = 1, 2, 3) v u Nm u 1 X RMSE(xi,k ) = t (xn − x ˆni,k|k )2 Nm n=1 i,k v u Nm X N u 1 X ARMSE(xi,k ) = t (xni,k − x ˆni,k|k )2 , Nm N n=1

(33)

k=1

where x ˆni,k|k and xni,k are the estimated and true state component respectively in the n-th experiment.

Jo

250

(32)

To let the comparison be fair, Nm = 5000 Monte Carlo runs are executed with simulation steps q N = 100. Divergence is claimed when estimation error (xn3,k − x ˆn3,k|k )2 exceeds 20 at arbitrarily discrete time k ∈ [0, N ].

Figure 2 shows a representative RMSE results of the existing methods and the proposed de20

Journal Pre-proof

255

correlating approaches with Sk = 0.5 and J = 2. The RMSE results of state component x1,k are similar to those of state component x2,k and omitted for convenience. The estimation performances w.r.t different iteration numbers J (J = 1, 2, . . . , 5) and different correlation parameters Sk

pro of

(Sk =0.1,0.2,. . . ,1) are illustrated as Figure 3 and Figure 4, respectively.

From Figures 2–4, we can see that performing iterations within a certain range (including the 260

iterations J and correlation term Sk ) can reduce the errors of the de-correlating IPLF and RTSIPLS. The de-correlating IPLF generally obtain a better performance than the other algorithms after 2–3 iterations with an acceptable average computational time, which can achieve an efficient balance between the estimated precision and calculating cost. Furthermore, it is much more robust than the others to the varying correlation parameters Sk . The iterated algorithms including the

265

de-correlating IPLF and RTS-IPLS exhibit varying degrees of performance loss with the increase

re-

number of iterations when the value of iterations J is large than 2 and the main reason is the numerical error. In other words, when the estimation error is significantly reduced, more iterations have little effect on improving the estimation accuracy and more numerical errors will be introduced. It is remarkable that the performance of the de-correlating PLS (RTS-IPLS) is not always more accurate than that of its corresponding de-correlating PLF (IPLF) approach for some applications

lP

270

due to the cross-correlated noises.

4.2. Univariate nonstationary growth model with one-step cross-correlated noises Because of the highly nonlinearities, the univariate nonstationary growth model [24, 34] is

urn a

frequently used in validating the performances of nonlinear estimation approaches, whose system equations are denoted as

xk+1 = 0.9xk + zk+1 =

10xk + 8 cos(1.2 ∗ k) + wk 1 + x2k

x3k + vk+1 . 20

(34) (35)

ˆ 0|0 = 1, and P0|0 = 4. System noises wk−1 and vk are The initial simulation parameters x0 = 5, x

275

Jo

correlated possessing cross-covariance Sk and their covariances are Q = 1 and R = 4, respectively. The RMSE and ARMSE averaged on all Monte Carlos runs are adopted as performance metrics to assess the estimation properties. To let the comparison be fair, Nm = 5000 Monte Carlo runs are implemented with simulation steps N = 50. Divergence is claimed when estimation error q ˆ nk|k )2 exceeds 30 at random discrete time k ∈ [0, N ]. (xnk − x 21

Journal Pre-proof

2.5 2 1.5 1 0.5 0 0

IPLF GASF GASS FB-IPLS RTS-IPLS

3

RMSE of x

3

RMSE of x

3.5

IPLF GASF GASS FB-IPLS RTS-IPLS

2.5 2 1.5

pro of

3.5

1

0.5

10

20

30

40

Time step k

50

(a)

0 0

10

20

30

40

Time step k

50

(b)

Figure 5: Performances of the considered algorithms for time step k with Sk = 0.5 and different iterations J. (a) RMSE in x with J = 2. (b) RMSE in x with J = 3. GASF

GASS

FB-IPLS

RTS-IPLS

1

0.6 1

2

3

lP

ARMSE of x

1.2

0.8

Number of iterations J

24

4

20 16

50

IPLF GASF GASS FB-IPLS RTS-IPLS

40 30 20

10 0 1

5

2

3

Number of iterations J

4

5

(b)

IPLF GASF GASS FB-IPLS RTS-IPLS

urn a

Average Computational Time (ms)

(a)

60

re-

IPLF

Number of Divergences

1.4

12

8

2

3

Number of iterations J

4

5

(c)

Jo

4 1

Figure 6: Performances of the considered algorithms for different iterations J with Sk = 0.5. (a) ARMSE in x. (b) Number of Divergences. (c) Average computational time.

22

Journal Pre-proof

1.5

120 GASS

FB-IPLS

RTS-IPLS

1.3 1.1 0.9 0.7 0.1

0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

0.8

0.9

1

(a)

100 80 60

IPLF GASF GASS FB-IPLS RTS-IPLS

pro of

GASF

Number of Divergences

ARMSE of x

IPLF

40 20

0 0.1

0.2

0.3

0.4

0.5

0.6

0.7

Correlation parameters Sk

0.8

0.9

1

(b)

Figure 7: Performances of the considered algorithms for different correlation parameters Sk with

re-

J = 2. (a) ARMSE in x. (b) Number of Divergences.

Figure 5 demonstrates two representative RMSE results of all considered algorithms with Sk = 280

0.5 and different iterations. The estimation performances w.r.t different iteration numbers J and different correlation parameters Sk are displayed as Figure 6 and Figure 7, respectively.

lP

Figures 5–7 show that the errors of the proposed iterated methods can be reduced by performing iterations and these approaches can obtain better performances than the compared algorithms after 2–3 iterations with acceptable computational cost. More precisely, except for a reasonable 285

computational complexity, the proposed IPLF (FB-IPLS or RTS-IPLS) method has advantages in estimation accuracy and robustness when compared with the existing approach GASF (GASS). An

urn a

effective balance between estimation accuracy and computational complexity can be obtained by the proposed iterated algorithms. It is worthy of note that the approximated RTS-IPLS algorithm has similar performances as the FB-IPLS approach except for the less computational cost. The reason 290

for this is that the effect of the noise correlation can be diminished by the iteration operations and then these two methods are equivalent with each other approximatively.

5. Conclusion

Jo

In the paper, several novel iterated posterior linearization filtering and smoothing approaches are introduced to improve the estimation performance for nonlinear models possessing cross-correlated 295

noises. In each iteration step of the proposed algorithms, SLRs of the system functions are performed w.r.t current posterior distribution and the multidimensional integrals of these SLRs are 23

Journal Pre-proof

approximated by some efficient sigma-point approaches. Simulation results demonstrate that the presented iterated methods not only have acceptable computational cost but also are more accurate

300

6. Future recommendation

pro of

and robust than the conventional approaches when the posterior PDF is unimodal.

Note that this paper only demonstrates the estimated performance of the presented techniques and their stabilities still need to be further proved, which is very important for the choice in practical applications. Furthermore, similar estimation approaches are able to extended for the continuous-discrete models with cross-correlated noises where the dynamic equations are expressed by stochastic differential equations (SDEs) [35, 36].

re-

305

7. Acknowledgment

The work was supported in part by the National Natural Science Foundation of China (Grant

8. References 310

lP

No. 61971100).

[1] B. Cui, X. Chen, X. Tang, H. Huang, X. Liu, Robust cubature kalman filter for GNSS/INS with missing observations and colored measurement noise, ISA Trans. 72 (2018) 138–146.

urn a

doi:doi.org/10.1016/j.isatra.2017.09.019. [2] M. Xiao, Y. Zhang, Z. Wang, H. Fu, An adaptive three-stage extended kalman filter for nonlinear discrete time system in presence of unknown inputs, ISA Trans. 75 (2018) 101–117. 315

doi:doi.org/10.1016/j.isatra.2018.02.007. [3] Y. Ho, R. Lee, A bayesian approach to problems in stochastic estimation and control, IEEE Trans. Autom. Control 9 (4) (1964) 333–339. doi:10.1109/TAC.1964.1105763.

Jo

[4] B. D. Anderson, J. B. Moore, Optimal filtering, Courier Corporation, Chelmsford, America, 2012. 320

[5] S. Julier, J. Uhlmann, Unscented filtering and nonlinear estimation, Proc. IEEE 92 (3) (2004) 401–422. doi:10.1109/JPROC.2003.823141. 24

Journal Pre-proof

[6] I. Arasaratnam, S. Haykin, Cubature Kalman filters, IEEE Trans. Autom. Control 54 (6) (2009) 1254–1269. doi:10.1109/TAC.2009.2019800.

325

pro of

[7] K. Ito, K. Xiong, Gaussian filters for nonlinear filtering problems, IEEE Trans. Autom. Control 45 (5) (2000) 910–927. doi:10.1109/9.855552.

[8] I. Arasaratnam, S. Haykin, Cubature Kalman smoothers, Automatica 47 (10) (2011) 2245– 2250. doi:10.1016/j.automatica.2011.08.005.

¨ ¨ Unscented Rauch–Tung–Striebel smoother, IEEE Trans. Autom. Control 53 (3) [9] S. SArkk A, (2008) 845–849. doi:10.1109/TAC.2008.919531. 330

[10] H. Lin, S. Sun, Distributed fusion estimator for multisensor multirate systems with correlated

re-

noises, IEEE Trans. Syst., Man, Cybern: Syst. 48 (7) (2018) 1131–1139. doi:10.1109/TSMC. 2016.2645599.

[11] C. Yang, Z. Yang, Z. Deng, Robust weighted state fusion kalman estimators for networked

335

2018.01.014.

lP

systems with mixed uncertainties, Inf. Fusion 45 (2019) 246–265. doi:10.1016/j.inffus.

´ [12] R. Caballero-Aguila, A. Hermoso-Carazo, J. Linares-P´erez, Z. Wang, A new approach to distributed fusion filtering for networked systems with random parameter matrices and correlated

urn a

noises, Inf. Fusion 45 (2019) 324–332. doi:10.1016/j.inffus.2018.02.006. [13] T. Tian, S. Sun, H. Lin, Distributed fusion filter for multi-sensor systems with finite-step 340

correlated noises, Inf. Fusion 46 (2019) 128–140. doi:10.1016/j.inffus.2018.05.002. [14] Y. Bar-Shalom, X. R. Li, T. Kirubarajan, Estimation with applications to tracking and navigation: Theory algorithms and software, John Wiley & Sons, New York, America, 2004. [15] X. Wang, Y. Liang, Q. Pan, F. Yang, A Gaussian approximation recursive filter for nonlinear systems with correlated noises, Automatica 48 (9) (2012) 2290–2297. doi:10.1016/j. automatica.2012.06.035.

Jo

345

[16] Y. Huang, Y. Zhang, N. Li, Z. Shi, Gaussian filter for nonlinear systems with correlated noises at the same epoch, Automatica 60 (2015) 122–126. doi:10.1016/j.automatica.2015.06. 035.

25

Journal Pre-proof

[17] G. Chang, Comment on “A Gaussian approximation recursive filter for nonlinear systems with 350

correlated noises”, Automatica 50 (2) (2014) 655–656. doi:10.1016/j.automatica.2013.12.

pro of

018. [18] X. Wang, Y. Liang, Q. Pan, Z. Wang, General equivalence between two kinds of noisecorrelation filters, Automatica 50 (12) (2014) 3316–3318. doi:10.1016/j.automatica.2014. 10.040. 355

[19] R. G. Brown, P. Y. C. Hwang, Introduction to random signals and applied Kalman filtering: with MATLAB exercises, Wiley, New York, America, 1997.

[20] D. Simon, Optimal state estimation: Kalman, H∞, and nonlinear approaches, Wiley, New

re-

York, America, 2006.

[21] Y. Huang, Y. Zhang, N. Li, Z. Shi, Design of Gaussian approximate filter and smoother for 360

nonlinear systems with correlated noises at one epoch apart, Circuits Syst. Signal Process. 35 (11) (2016) 3981–4008. doi:10.1007/s00034-016-0256-0.

lP

[22] R. Zhan, J. Wan, Iterated unscented Kalman filter for passive target tracking, IEEE Trans. Aerosp. Electron. Syst. 43 (3) (2007) 1155–1163. doi:10.1109/TAES.2007.4383605. ´ F. Garc´ıa-Fern´andez, L. Svensson, M. R. Morelande, S. S¨ [23] A. arkk¨ a, Posterior linearization filter: 365

Principles and implementation using sigma points, IEEE Trans. Signal Process. 63 (20) (2015)

urn a

5561–5573. doi:10.1109/TSP.2015.2454485. ´ F. Garc´ıa-Fern´andez, L. Svensson, S. S¨ [24] A. arkk¨ a, Iterated posterior linearization smoother, IEEE Trans. Autom. Control 62 (4) (2017) 2056–2063. doi:10.1109/TAC.2016.2592681. [25] B. M. Bell, F. W. Cathey, The iterated Kalman filter update as a Gauss-Newton method, 370

IEEE Trans. Autom. Control 38 (2) (1993) 294–297. doi:10.1109/9.250476. [26] B. M. Bell, The iterated Kalman smoother as a Gauss-Newton method, SIAM J. Optim. 4 (3)

Jo

(1993) 626–636. doi:10.1137/0804035. [27] X. Hu, M. Bao, X.-P. Zhang, L. Guan, Y.-H. Hu, Generalized iterated Kalman filter and its performance evaluation, IEEE Trans. Signal Process. 63 (12) (2015) 3204–3217. doi: 375

10.1109/TSP.2015.2423266. 26

Journal Pre-proof

[28] A. Gelb, Applied optimal estimation, MIT Press, Cambridge, America, 1974. [29] I. Arasaratnam, S. Haykin, R. J. Elliott, Discrete-time nonlinear filtering algorithms using

pro of

Gauss-Hermite quadrature, IEEE Trans. Signal Process. 95 (5) (2007) 953–977. doi:10.1109/ JPROC.2007.894705. 380

[30] H. E. Rauch, F. Tung, C. T. Striebel, Maximum likelihood estimates of linear dynamic systems, AIAA J. 3 (8) (1965) 1445–1450. doi:doi.org/10.2514/3.3166.

[31] D. C. Frmer, J. E. Potter, The optimum linear smoother as a combination of two optimum linear filters, IEEE Trans. Autom. Control 14 (4) (1969) 387–390. doi:10.1109/TAC.1969. 1099196.

[32] F. Lewis, Optimal estimation with an introduction to stochastic control, John Wiley & Sons, New York, America, 1986.

re-

385

[33] R. Hunger, Floating point operations in matrix-vector calculus, Tech. Rep., Munich University

lP

of Technology, 2007.

[34] Y. Wu, D. Hu, M. Wu, X. Hu, A numerical-integration perspective on Gaussian filters, IEEE 390

Trans. Signal Process. 54 (8) (2006) 2910–2921. doi:10.1109/TSP.2006.875389. [35] Y. Wang, H. Zhang, X. Mao, Y. Li, Accurate smoothing methods for state estimation of

urn a

continuous-discrete nonlinear dynamic systems, IEEE Trans. Autom. Control 64 (10) (2019) 4284–4291. doi:10.1109/TAC.2019.2893876. [36] Y. Wang, H. Zhang, Accurate smoothing for continuous-discrete nonlinear systems with nongaussian noise, IEEE Signal Process. Lett. 26 (3) (2019) 465–469. doi:10.1109/LSP.2018. 2890313.

Jo

395

27

*Conflict of Interest

Journal Pre-proof

Conflict of interest statement We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work,

pro of

there is no professional or other personal interest of any nature or kind in any product, service and/or company that could be construed as

influencing the position presented in, or the review of, the manuscript

Jo

urn a

lP

re-

entitled.