Skip to main content
Entropy logoLink to Entropy
. 2020 Sep 3;22(9):982. doi: 10.3390/e22090982

A Novel Perspective of the Kalman Filter from the Rényi Entropy

Yarong Luo 1, Chi Guo 1,2,*, Shengyong You 1, Jingnan Liu 1,2
PMCID: PMC7597296  PMID: 33286750

Abstract

Rényi entropy as a generalization of the Shannon entropy allows for different averaging of probabilities of a control parameter α. This paper gives a new perspective of the Kalman filter from the Rényi entropy. Firstly, the Rényi entropy is employed to measure the uncertainty of the multivariate Gaussian probability density function. Then, we calculate the temporal derivative of the Rényi entropy of the Kalman filter’s mean square error matrix, which will be minimized to obtain the Kalman filter’s gain. Moreover, the continuous Kalman filter approaches a steady state when the temporal derivative of the Rényi entropy is equal to zero, which means that the Rényi entropy will keep stable. As the temporal derivative of the Rényi entropy is independent of parameter α and is the same as the temporal derivative of the Shannon entropy, the result is the same as for Shannon entropy. Finally, an example of an experiment of falling body tracking by radar using an unscented Kalman filter (UKF) in noisy conditions and a loosely coupled navigation experiment are performed to demonstrate the effectiveness of the conclusion.

Keywords: Rényi entropy, discrete Kalman filter, continuous Kalman filter, algebraic Riccati equation, nonlinear differential Riccati equation

1. Introduction

In the late 1940s, Shannon introduced a logarithmic measure of information [1] and a theory that included information entropy (the literature shows that it is related to Boltzmann entropy in statistical mechanics). The more stochastic and unpredictable a variable is, the larger its entropy is. As a measure of information, entropy has been used in various fields, such as information theory, signal processing, information-theoretic learning [2,3], etc. As a generalization of the Shannon entropy, Rényi entropy, named after Alfréd Rényi [4], allows for different averaging of probabilities through a control parameter α, and is usually used to quantify the diversity, uncertainty, or randomness of random variables. Liang [5] presented the evolutionary entropy equations and the uncertainty estimation for Shannon entropy and relative entropy, which is also called Kullback–Leibler divergence [6], within the framework of dynamical systems. However, higher-order Rényi entropy has some better properties than Shannon entropy by setting the control parameter α in most cases.

The Kalman filter [7] and its variants have been widely used in navigation, control, tracking, etc. Many works focus on combining different entropy and entropy-like quantities with the original Kalman filter to improve the performance. When the state space equation is nonlinear, Rényi entropy can be used to measure the nonlinearity [8,9]. Shannon entropy was used to estimate the weight of each particle from the weights of different measurement models for the fusion algorithm in [10]. Quadratic Rényi entropy [11] of innovation has been used as a minimum entropy criterion under a nonlinear and non-Gaussian circumstance [12] in unscented Kalman filter (UKF) [13] and finite mixtures [14]. A generalized density evolution equation [15] and polynomial-based non-linear compensation [16] were used to improve the minimum entropy filtering [17]. Relative entropy has been used to measure the similarity between the probabilistic density functions during the recursive processes of the nonlinear filter [18,19]. As for the nonlinear measurement equation with additive Gaussian noise, relative entropy can be deduced to measure the nonlinearity of the measurement [20], and can also be used to measure the approximation error of the i-th measurement element in the partitioned update Kalman filter [21]. When the state variables and the measurement variables do not belong to strict Gaussian distribution, such as in the seamless indoor/outdoor multi-source fusion positioning problem [22], the estimation error can be measured by the relative entropy. Relative entropy can also be used to calculate the number of particles in the unscented particle filter for mobile robot self-localization [23] and to calculate the sample window size in the cubature Kalman filter (CKF) [24] for attitude estimation [25]. Moreover, it has been verified that the original Kalman filter can be derived by maximizing the relative entropy [26]. Meanwhile, the robust maximum correntropy criterion has been adopted as the optimal criterion to derive the maximum correntropy Kalman filter [27,28]. However, there has been no work on the direct connections between the Rényi entropy and the Kalman filter theory until now.

In this paper, we propose a new perspective of the Kalman filter from the Rényi entropy for the first time, which bridges the gap between the Kalman filter and the Rényi entropy. We calculate the temporal derivative of the Rényi entropy for the Kalman filter mean square error matrix, which provides the optimal recursive solution mathematically and will be minimized to obtain the Kalman filter gain. Moreover, from the physical point of view, the continuous Kalman filter approaches a steady state when the temporal derivative of the Rényi entropy is equal to zero, which also means that the Rényi entropy will keep stable. A numerical experiment of falling body tracking in noisy conditions with radar using the UKF and a practical experiment of loosely-coupled integration are provided to demonstrate the effectiveness of the above conclusion.

The structure of this paper is as follows. In Section II, the definitions and properties of Shannon entropy and Rényi entropy are presented. In Section III, the Kalman filter is derived from the perspective of minimizing the temporal derivative of Rényi entropy, and the connection between the Rényi entropy and the algebraic Riccati equation is explained. In Section IV, experimental results and analysis are given by the simulation of the UKF and the real integrated navigation data. We finally conclude this paper and provide an outlook for future work in Section V.

2. The Connection between the Kalman Filter and the Temporal Derivative of the Rényi Entropy

2.1. Rényi Entropy

To calculate the Rényi entropy of the continuous probability density function (PDF), it is necessary to extend the definition of the Rényi entropy to the continuous form. The Rényi entropy of order α for a continuous random variable with a multivariate Gaussian PDF p(x) is defined [4] and calculated [9] as:

HRα(x)=11αlnSpα(x)dx=N2ln(2πα1α1)+12ln(detΣ), (1)

where α>0,α1, and α is a parameter providing a family of entropy functions. N is the dimension of the random variable x. S is the support. Σ is the covariance matrix of p(x).

It is straightforward to show that the temporal derivative of the Rényi entropy is given by [9]:

H˙R(α)(x)=12Tr{Σ1Σ˙}, (2)

where Σ˙ is the temporal derivative of the covariance matrix and Tr(·) is the trace operator.

It is easy to get the Shannon entropy for the multivariate Gaussian PDF by taking the limitation of Equation (1) as α approaches 1. This entropy is given as H(x)=N2ln(2πe)+12ln(detΣ), and the temporal derivative of the Shannon entropy is given as H˙(x)=12Tr{Σ1Σ˙}. It is obvious the temporal of the Shannon entropy is the same as the temporal of the Rényi entropy. Therefore, we will see later that the conclusion can also be derived from the temporal derivative of the Shannon entropy. However, the Rényi entropy for the multivariate Gaussian PDF instead of the temporal derivative of the Rényi entropy will be used by adjusting the free parameter α for different uncertainty measurements in most cases, as the filtering problem has to account for the nonlinearity and the non-Gaussian noise; we adopt the Rényi entropy as the measurement for uncertainty.

2.2. Kalman Filter

Given the continuous-time linear system [29]:

X˙(t)=F(t)X(t)+G(t)w(t) (3)
Z(t)=H(t)X(t)+v(t), (4)

where X(t) is the state vector; F(t) is the state transition matrix; G(t) is the system noise driving matrix; Z(t) is the measurement vector; H(t) is the measurement matrix; and w(t) and v(t) are independent white Gaussian noise with zero mean value; their covariance matrices are Q(t) and R(t), respectively:

E[w(t)]=0,E[w(t)wT(τ)]=Q(t)δ(tτ) (5)
E[v(t)]=0,E[v(t)vT(τ)]=R(t)δ(tτ) (6)
E[w(t)vT(τ)]=0, (7)

where δ(t) is the Dirac impulse function, Q(t) is a symmetric non-negative definite matrix, and R(t) is a symmetric positive matrix.

The continuous Kalman filter can be deduced by taking the limit of the discrete Kalman filter. The discrete-time state-space model is arranged as follows [29]:

Xk=Φk|k1Xk1+Γk|k1Wk1 (8)
Zk=HkXk+Vk (9)

where Xk is an n-dimensional state vector; Zk is an m-dimensional measurement vector; Φk|k1, Γk|k1, and Hk are the known system structure parameters, which are called the n×n dimensional one-step state update matrix, the n×l dimensional system noise distribution matrix, and the m×n dimensional measurement matrix, respectively; Wk1 is the l-dimensional system noise vector, and Vk is the m-dimensional measurement noise vector. Both of them are Gaussian noise vector sequences with zero mean value, and are independent of each other:

E[Wk]=0,E[WkWjT]=Qkδkj (10)
E[Vk]=0,E[VkVjT]=Rkδkj (11)
E[WkVjT]=0. (12)

The above equation is the basic assumption for the noise requirement in the Kalman filtering state space model, where Qk is a symmetric non-negative definite matrix, and Rk is a symmetric positive definite matrix. δkj is the Kronecker δ function.

The covariance parameters Qk and Rk play roles similar to those of Q and R in the continuous filter, but they do not have the same numerical values. Next, the relationship between the corresponding continuous and discrete filter parameters will be derived.

To achieve the transformation from the continuous form to the discrete form, the relations between Q and R and the corresponding Qk and Rk for a small step size Ts are needed. According to the linear system theory, the relation between Q and Qk from Equation (3) to Equation (8) is as follows:

Φk|k1=Φ(tk,tk1)etk1tkF(τ)dτ (13)
Γk|k1Wk1=tk1tkΦ(tk,τ)G(τ)w(t)dτ. (14)

Denote the discrete-time interval as Ts=tktk1, when F(t) does not change too dramatically within the shorter integral interval [tk1,tk]. Take the Taylor expansion of eF(tk1)Ts with respect to F(tk1)Ts and set F(tk1)Ts<<I, so the higher-order terms are negligible and the one-step transition matrix, Equation (13), can be approximated as:

Φk|k1eF(tk1)Ts=I+F(tk1)Ts+F2(tk1)Ts22!+F3(tk1)Ts33!+I+F(tk1)Ts. (15)

Equation (14) shows that Γk|k1Wk1 is the linear transform of the Gaussian white noise w(τ); the result remains the normal distribution random vector. Therefore, the first- and second-order statistical characteristics can be used to describe and be equivalent to Γk|k1Wk1. Referring to Equation (5), the mean of Γk|k1Wk1 is given as follows:

E[Γk|k1Wk1]=E[tk1tkΦ(tk,τ)G(τ)w(τ)dτ]=tk1tkΦ(tk,τ)G(τ)E[w(τ)]dτ=0. (16)

For the second-order statistical characteristics, when kj, the time parameter between the noise w(τk) and w(τj) is independent, so Γk|k1Wk1 and Γj|j1Wj1 are uncorrelated:

E[(Γk|k1Wk1)(Γj|j1Wj1)T]=0(kj). (17)

When k=j, thus

E[(Γk|k1Wk1)(Γk|k1Wk1)T]=E[tk1tkΦ(tk,τ)G(τ)w(τ)dτ][tk1tkΦ(tk,s)G(s)w(s)ds]T=Etk1tkΦ(tk,τ)G(τ)w(τ)tk1tkwT(s)GT(s)ΦT(tk,s)dsdτ=tk1tkΦ(tk,τ)G(τ)tk1tkE[w(τ)wT(s)]GT(s)ΦT(tk,s)dsdτ. (18)

Substituting Equation (5) into the above equation:

E[(Γk|k1Wk1)(Γk|k1Wk1)T]=tk1tkΦ(tk,τ)G(τ)tk1tkQ(t)δ(τs)GT(s)ΦT(tk,s)dsdτ=tk1tkΦ(tk,τ)G(τ)Q(τ)GT(τ)ΦT(tk,τ)dτ. (19)

When the noise control matrix G(τ) changes slowly during the time interval [tk1,tk], Equation (19) becomes:

E[(Γk|k1Wk1)(Γk|k1Wk1)T]tk1tk[I+F(tk1)(tkτ)]G(tk1)Q(τ)GT(tk1)[I+F(tk1)(tkτ)]Tdτ=[I+F(tk1)Ts]·[G(tk1)Q(tk1)GT(tk1)Ts]·[I+F(tk1)Ts]T+112F(tk1)G(tk1)Q(tk1)GT(tk1)F(tk1)TTs3[I+F(tk1)Ts]G(tk1)·[Q(tk1)Ts]·[I+F(tk1)Ts]G(tk1)T. (20)

When F(tk1)Ts<<I is satisfied, the above equation can be further approximated:

E[(Γk|k1Wk1)(Γk|k1Wk1)T]G(tk1)·[Q(tk1)Ts]·GT(tk1). (21)

Comparing the result with Equation (10):

Γk|k1[I+F(tk1)Ts]G(tk1)G(tk1) (22)
E[WkWjT]=Qkδkj=[Q(tk)Ts]δkj. (23)

Notice that [29]:

Qk=Q(tk)Ts. (24)

The derivation of the equation relating to Rk and R is more subtle. In the continuous model, v(t) is white, so simple sampling of Z(t) leads to measurement noise with infinite variance. Hence, in the sampling process, we have to imagine averaging the continuous measurement over the Ts interval to get an equivalent discrete sample. This is justified because x is not the Gaussian white noise and can be approximately constant within the interval.

Zk=1Tstk1tkZ(t)dt=1Tstk1tk[H(t)x(t)+v(t)]dt=H(tk)xk+1Tstk1tkv(t)dt. (25)

Then, the discrete noise matrix and the continuous noise matrix are equivalent:

Vk=1Tstk1tkv(t)dt. (26)

From Equation (12), we have:

E[VkVjT]=Rkδkj=1Ts2tk1tktj1tjE[v(τ)v(s)]dτds=1Ts2tk1tktj1tjR(τ)δ(sτ)dτds=1Ts2tk1tkR(τ)δkjdτR(tk)Tsδkj. (27)

Comparing it with Equation (6), we have [29]:

Rk=R(tk)Ts. (28)

2.3. Derivation of the Kalman Filter

Assuming that the optimal state estimation at tk1 is X^k1, the state estimation error is X˜k1, and the state estimation covariance matrix is Σk1:

X˜k1=Xk1X^k1 (29)

and

Σk1=E[X˜k1X˜k1T]=E[(Xk1X^k1)(Xk1X^k1)T]. (30)

If we take the expectation operator of both sides of Equation (8), we obtain the state one-step prediction and the state one-step estimation error:

Xk|k1=E[Xk]=E[Φk|k1Xk1+Γk|k1Wk1]=Φk|k1E[Xk1]=Φk|k1X^k1, (31)
X˜k|k1=XkXk|k1. (32)

Substituting Equations (8) and (31) into Equation (32) leads to:

X˜k|k1=(Φk|k1Xk1+Γk|k1Wk1)Φk|k1X^k1=Φk|k1(Xk1X^k1)+Γk|k1Wk1=Φk|k1X˜k1+Γk|k1Wk1. (33)

Since X˜k1 is uncorrelated with Wk1, we therefore obtain the covariance of the state one-step estimation error X˜k|k1 as follows:

Σk|k1=E[X˜k|k1X˜k|k1T]=E[(Φk|k1X˜k1+Γk|k1Wk1)(Φk|k1X˜k1+Γk|k1Wk1)T]=Φk|k1E[X˜k1X˜k1T]Φk|k1T+Γk|k1E[Wk1Wk1T]Γk|k1T=Φk|k1Σk1Φk|k1T+Γk|k1Qk1Γk|k1T. (34)

In a similar way, the measurement at tk can be predicted by the state one-step estimation prediction Xk|k1 and system measurement Equation (9) as follows:

Zk|k1=E[HkXk|k1+Vk]=HkXk|k1. (35)

In fact, there is difference between the measurement one-step prediction Zk|k1 and the actual measurement Zk. The difference is denoted as measurement one-step prediction error:

Z˜k|k1=ZkZk|k1. (36)

Substituting the measurement Equations (9) and (35) into Equation (36) yields:

Z˜k|k1=ZkHkXk|k1=HkXk+VkHkXk|k1=HkX˜k|k1+Vk. (37)

In general, the measurement one-step prediction error Z˜k|k1 is called innovation in the classical Kalman filter theory, and it indicates the new information about the state estimate carried by the measurement one-step prediction error.

On the one hand, if the estimation of Xk only includes the state one-step prediction Xk|k1 of the system state equation, the estimation accuracy will be low, as no information of the measurement equation has been used. On the other hand, according to Equation (37), the measurement one-step prediction error calculated using the system measurement equation contains the information of the state one-step prediction of Xk|k1. Consequently, it is natural to consider all the state information that comes from the system state equation and the measurement equation, respectively, and correct the state one-step prediction mean Xk|k1 with the measurement one-step prediction error Z˜k|k1. Thereby, the optimal estimation of Xk can be calculated by the combination of Xk|k1 and Z˜k|k1 as follows:

X^k=Xk|k1+KkZ˜k|k1, (38)

where Kk is the undetermined correction factor matrix.

Substituting Equations (31) and (37) into Equation (38) obtains:

X^k=Xk|k1+Kk(ZkHkXk|k1)=(IKkHk)Xk|k1+KkZk=(IKkHk)Φk|k1X^k1+KkZk. (39)

From Equation (39), the current state estimation X^k is a linear combination of the last state estimation X^k1 and the current measurement Zk, which considers the influence of the structural parameters Φk|k1 in the state equation and the structure parameters Hk in the measurement equation with different types of construction.

The state estimation error at the current time tk is denoted as:

X˜k=XkX^k, (40)

where Xk is the true values and X^k is the posterior estimation of Xk.

Substituting Equation (39) into Equation (40) obtains:

X˜k=Xk[Xk|k1+Kk(ZkHkXk|k1)]=(XkXk|k1)Kk(HkXk+VkHkXk|k1)=X˜k|k1Kk(HkX˜k|k1+Vk)=(IKkHk)X˜k|k1KkVk. (41)

Then, the mean square error matrix of state estimation X^k is given by:

Σk=E[X˜kX˜kT]=E{[(IKkHk)X˜k|k1KkVk][(IKkHk)X˜k|k1KkVk]T}=(IKkHk)E[X˜k|k1X˜k|k1T](IKkHk)T+KkE[VkVkT]KkT=(IKkHk)Σk|k1(IKkHk)T+KkRkKkT. (42)

Substituting Equation (34) into Equation (42) obtains:

Σk=(IKkHk)[Φk|k1Σk1Φk|k1T+Γk|k1Qk1Γk|k1T](IKkHk)T+KkRkKkT=Φk|k1Σk1Φk|k1T+KkHkΦk|k1Σk1Φk|k1THkTKkTΦk|k1Σk1Φk|k1THkTKkTKkHkΦk|k1Σk1Φk|k1T+Γk|k1Qk1Γk|k1TKkHkΓk|k1Qk1Γk|k1TΓk|k1Qk1Γk|k1THkTKkT+KkHkΓk|k1Qk1Γk|k1THkTKkT+KkRkKkT. (43)

We now use the approximation Φk|k1I+F(tk1)Ts as Equation (15). From Equation (22) with Γk|k1G(tk1), we have:

Σk=[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T]+KkHk[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTKkT[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTKkTKkHk[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T+G(tk1)Qk1GT(tk1)KkHkG(tk1)Qk1GT(tk1)G(tk1)Qk1GT(tk1)HkTKkT+KkHkG(tk1)Qk1GT(tk1)HkTKkT+KkRkKkT. (44)

Note from Equation (24) that Qk is of the order of Ts and from Equation (28) that Rk=R(tk)Ts; then, Equation (44) becomes:

Σk=[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T]+KkHk[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTKkT[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTKkTKkHk[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T+G(tk1)Q(tk)TsGT(tk1)KkHkG(tk1)Q(tk)TsGT(tk1)G(tk1)Q(tk)TsGT(tk1)HkTKkT+KkHkG(tk1)Q(tk)TsGT(tk1)HkTKkT+KkR(tk)TsKkT. (45)

2.4. The Temporal Derivative of the Rényi Entropy and the Kalman Filter Gain

To obtain the continuous form of covariance matrix Σ, the limit will be taken. However, the relation between the undetermined correction factor matrix Kk and its continuous form still remains unknown. Therefore, we make the following assumption.

Assumption 1.

Kk is of the order of Ts, that is:

K(tk)=KkTs. (46)

From the conclusion, we can also derive this assumption conversely. We next draw the conclusion as one theorem under the assumption, as follows:

Theorem 1.

The discrete form of the undetermined correction factor matrix is the same as the continuous form when the temporal derivative of Rényi entropy is minimized. This can be presented in a mathematical form as follows:

{Kk=ΣkHkTRk,K=ΣHTR1|K*=argminKH˙R(α)(K)}. (47)

Proof of Theorem 1.

We substitute the expression for Kk into Equation (45) and neglect higher-order terms in Ts; Equation (45) becomes:

Σk=[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T]+TsK(tk)HkΣk1[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTTsKT(tk)[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]THkTTsKT(tk)TsK(tk)Hk[I+F(tk1)Ts]Σk1[I+F(tk1)Ts]T+G(tk1)Q(tk)TsGT(tk1)TsK(tk)HkG(tk1)Q(tk)TsGT(tk1)G(tk1)Q(tk)TsGT(tk1)HkTTsKT(tk)+TsK(tk)HkG(tk1)Q(tk)TsGT(tk1)HkTTsKT(tk)+TsK(tk)RkTsTsKT(tk)=Σk1+TsF(tk1)Σk1+TsΣk1FT(tk1)Σk1HkTTsK(tk)TTsK(tk)HkΣk1+G(tk1)Q(tk)TsGT(tk1)+TsK(tk)R(tk)TsTsKT(tk). (48)

Moving the first term of Equation (48) from right to left and dividing both sides by Ts to form the finite difference expression:

ΣkΣk1Ts=F(tk1)Σk1+Σk1FT(tk1)Σk1HkTK(tk)TK(tk)HkΣk1+G(tk1)Q(tk)GT(tk1)+K(tk)R(tk)KT(tk). (49)

Finally, passing to the limit as Ts0 and dropping of the subscripts lead to the matrix differential equation:

Σ˙=FΣ+ΣFTΣHTKTKHΣ+GQGT+KRKT. (50)

Σ is invertible, as it is a positive matrix. Multiplying Σ1 with Equation (50), we can consider the temporal derivative of the Rényi entropy of the mean square error matrix Σ using Equation (2):

H˙R(α)=12Tr{Σ1Σ˙}=12Tr{Σ1FΣ+FTHTKTΣ1KHΣ+Σ1GQGT+Σ1KRKT}=12Tr{F+FTHTKTKH+Σ1GQGT+Σ1KRKT}=12Tr{2F2KH+Σ1GQGT+Σ1KRKT}, (51)

where the invariance under the cyclic permutation property of the trace operator has been used to eliminate Σ1 and Σ, as well as the truth that Tr(F)=Tr(FT) has been used to simplify the formula.

It is obvious that Equation (51) is a quadratic function of the undetermined correction factor matrix K. Thereby, there must be a minimum of H˙R(α)(x) in a probabilistic sense. Taking the derivative of both sides of Equation (51) with respect to matrix K obtains:

KH˙R(α)=2Tr(KH)K+Tr(Σ1KRKT)K=2HT+Tr(Σ1KR(K)T)K+Tr(Σ1(K)RKT)K=2HT+Σ1KR+(RKTΣ1)T. (52)

In addition, since Σ1 and Rk are symmetric matrices, the result is:

KH˙R(α)=2HT+2Σ1KR. (53)

Rk is invertible, as it is a positive matrix. According to the extreme value principle of the function, when the above are equal to zero, then we have:

K=ΣHTR1. (54)

So far, we have found the analytic solution to the undetermined correction factor matrix K, which is called the continuous-time Kalman filter gain in the classical Kalman filter. Then, the recursive formulations of the Kalman filter can be established through the Kalman filter gain K. Most importantly, this implies the connection between the temporal derivative of Rényi entropy and the classical Kalman filter: The temporal derivative of the Rényi entropy is minimized when the Kalman filter gain satisfies Equation (54).

Looking back to Assumption 1 and substituting Equation (28) into Equation (54), we obtain:

K(tk)=KkTs=K=ΣHTR1=ΣkHkTRk(Ts)=ΣkHkTRkTs. (55)

Therefore, the discrete-time Kalman filter gain can be expressed as follows:

Kk=ΣkHkTRk. (56)

Remark 1.

The discrete-time Kalman filter gain has the same form as the continuous-time filter gain, as shown in the Equation (54). In principle, this is consistent with our intuition and proves the correctness and rationality of Assumption A1, in turn.

Remark 2.

The Kalman filter gain is equivalent to the minimization of the temporal derivative of the Rényi entropy, although it has the same result as the original Kalman filter, which is deduced under the minimum mean square error criterion.

Substituting Equation (54) into Equation (50), we have:

Σ˙=FΣ+ΣFTΣHTKTΣHTR1HΣ+GQGT+ΣHTR1RKT=FΣ+ΣFTΣHTR1HΣ+GQGT. (57)

This is a second-order nonlinear differential equation with respect to the mean square error matrix Σ, and it is commonly called the Riccati equation. This is the same result as that of the Bucy–Kalman filter [7].

If the system equation, Equation (3), and the measurement equation, Equation (4), form a linear time-invariant system with constant noise covariance, the mean square error matrix Σ may reach a steady-state value, and Σ˙ may eventually reach zero. So, we have the continuous algebraic Riccati equation as follows:

Σ˙=FΣ+ΣFTΣHTR1HΣ+GQGT=0. (58)

As we can see, the time derivative of covariance at the steady state is zero; then, the temporal derivative of the Rényi entropy should also be zero:

H˙R(α)=0. (59)

This implies that when the system approaches a stable state, the Rényi entropy approaches a steady value so that the temporal derivative of the Rényi entropy is zero. This is reasonable when the steady system owns a constant Rényi entropy, as uncertainty is stable, which follows our intuitive understanding. Consequently, it is worth noting that whether the value of the Rényi entropy is stable or not can be a validated indicator of whether the system is approaching the steady state.

3. Simulations and Analysis

In this section, we give two experiments to show that when the nonlinear filter system approaches the steady state, the Rényi entropy of the system approaches stability. The first experiment is a numerical example of a falling body in noisy conditions, tracked by radar [30] using the UKF. The second experiment is a practical experiment of loosely coupled integration [29]. The simulations were carried out on MATLAB 2018a running on a computer with i5-5200U, 2.20 GHz CPU, and the graphs were plotted by MATLAB.

3.1. Falling Body Tracking

In the example of a falling body being tracked by radar, the body falls vertically. The radar is placed at a vertical distance L from the body, and the radar measures the distance y from the radar to the body. The state-space equation of the body is given by:

x˙1=x2x˙2=d+gx˙3=0, (60)

where x1 is the height, x2 is the velocity, x3 is the ballistic coefficient, g=9.81 m/s2 is the gravity acceleration, and d is the air drag, which could be approximated as:

d=ρx222x3=ρ0exp(x1k)x222x3, (61)

where ρ is the air density with an initial value of ρ0=1.225; ρ0=1.225 and k=6705.6 are constants.

The measurement equation is:

y=L2+x12. (62)

It is worth noting that the drag and the square root cause severely nonlinearity in the state-space function and measurement function, respectively.

The discrete-time nonlinear system can be given by the Euler discretization method. Combining the additive process with Gaussian white noises for measurement, we can obtain:

x1(n+1)=x1(n)+x2(n)·T+w1(n)x2(n+1)=x2(n)+(d+g)·T+w2(n)x3(n+1)=x3(n)+w3(n) (63)
y(n)=L2+x12(n)+v(n). (64)

In the UKF numerical experiment, we set the sampling period to T=0.4 s, the horizontal distance to L=100 m, the maximum number of samples to N=100, the process noise to Sw=diag(105,103,102), the measurement noise to Sv=106, and the initial state to x=[105;5000;400]. The results are shown as follows:

Figure 1 shows the evolution of covariance matrix Σ. Figure 2 and Figure 3 show the Rényi entropy of covariance matrix Σ and its change in adjacent time, respectively. Notice that the uncertainty increases near the middle of the plots, which is coincident with the drag peak. However, the Rényi entropy fluctuates around 15; even the fourth element of Σ changes dramatically. Of course, the entropy changes are closely accompanied by the drag peak, which means the change of the entropy of covariance reflects the evolution of matrix Σ. Consequently, the Rényi entropy can be viewed as the indicator of whether the system is approaching the steady state or not.

Figure 1.

Figure 1

Evolution of matrix Σ.

Figure 2.

Figure 2

Simulation results for the entropy.

Figure 3.

Figure 3

Simulation results for the change of entropy.

3.2. Practical Integrated Navigation

In the loosely integrated navigation system, the system state parameter x is composed of inertial navigation system (INS) error states in the North–East–Down (NED) local-level navigation frame, and can be expressed as follows:

x=[(δrn)T(δvn)T(ψ)T(bg)T(ba)T]T, (65)

where δrn, δvn, and ψ represent the position error, the velocity error, and the attitude error, respectively; bg and ba are modeled as first-order Gauss–Markov processes, representing the gyroscope bias and the accelerometer bias, respectively.

The discrete-time state update equation is used to update state parameters as follows:

xk=Φk|k1xk1+Gk|k1wk1, (66)

where Gk|k1 is the system noise matrix, wk1 is the system noise, and Φk|k1 is the state transition matrix from tk1 to tk; this is determined by the dynamic model of the state parameter.

In the loosely coupled integration, the measurement equation can be simply expressed as:

δz=Hkxk+vk, (67)

where vk is the measurement noise, Hk is the measurement matrix, and zk is the measurement vector calculated by subtracting the global navigation satellite system (GNSS) observation with the inertial navigation system (INS) mechanism.

The experiments reported in this section were carried out by processing the data from an unmanned ground vehicle test. The gyroscope random walk was set to 0.03 deg/h and the velocity random walk was set to 0.05 m/s/h. The sampling rates of the inertial measurement unit (IMU) and the GNSS are 200 Hz and 1 Hz, respectively. The test lasts 48 min.

The position error curve, velocity error curve, and attitude error curve of the loosely coupled integration are shown in Figure 4, Figure 5 and Figure 6. The root mean squares (RMSs) of the position errors in the north, east, and earth directions are 0.0057 m, 0.0024 m, and 0.0134 m, respectively. The RMS of the velocity errors in the north, east, and earth directions are 0.0023 m/s, 0.0021 m/s, and 0.0038 m/s, respectively. The RMSs of the attitude errors in the roll, pitch, and yaw directions are 0.0034 deg, 0.0030 deg, and 0.0178 deg, respectively.

Figure 4.

Figure 4

Position error of the loosely coupled integration.

Figure 5.

Figure 5

Velocity error of the loosely coupled integration.

Figure 6.

Figure 6

Attitude error of the loosely coupled integration.

The Rényi entropy of the covariance P is shown in Figure 7. As we can see, the Rényi entropy fluctuates around 100 once the filter converges, which is consistent with the conclusion from the entropy perspective.

Figure 7.

Figure 7

Rényi entropy of the covariance Σ.

4. Conclusions and Final Remarks

We have considered the original Kalman filter by taking the minimization of the temporal derivative of the Rényi entropy. In particular, we show that the temporal derivative of Rényi entropy is equal to zero when the Kalman filter system approaches the steady state, which means that the Rényi entropy approaches a stable value. Finally, simulation experiments and practical experiments show the Rényi entropy truly stays stable when the system becomes steady.

Future work includes calculating the Rényi entropy of the innovation term when the measurements and the noise are non-Gaussian [14] in order to evaluate the effectiveness of measurements and adjust the noise covariance matrix. Meanwhile, we can also calculate the Rényi entropy of the nonlinear dynamical equation to measure the nonlinearity in the propagation step.

Acknowledgments

In this section you can acknowledge any support given which is not covered by the author contribution or funding sections. This may include administrative and technical support, or donations in kind (e.g., materials used for experiments).

Author Contributions

Conceptualization, Y.L. and C.G.; Funding acquisition, C.G. and J.L.; Investigation, Y.L.; Methodology, Y.L., C.G., and S.Y.; Project administration, J.L.; Resources, C.G.; Software, Y.L. and S.Y.; Supervision, J.L.; Validation, S.Y.; Visualization, S.Y.; Writing—original draft, Y.L.; Writing—review and editing, C.G., S.Y., and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant from the National Key Research and Development Program of China (2018YFB1305001).

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Shannon C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948;27:379–423. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
  • 2.Principe J.C. Information Theoretic Learning: Renyi’s Entropy and Kernel Perspectives. Springer Science & Business Media; Berlin, Germany: 2010. [Google Scholar]
  • 3.He R., Hu B., Yuan X., Wang L. Robust Recognition via Information Theoretic Learning. Springer International Publishing; Berlin, Germany: 2014. [Google Scholar]
  • 4.Rényi A. On measures of entropy and information; Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability; Berkeley, CA, USA. 20 June–30 July 1961. [Google Scholar]
  • 5.Liang X.S. Entropy evolution and uncertainty estimation with dynamical systems. Entropy. 2014;16:3605–3634. doi: 10.3390/e16073605. [DOI] [Google Scholar]
  • 6.Kullback S., Leibler R.A. On information and sufficiency. Ann. Math. Stat. 1951;22:79–86. doi: 10.1214/aoms/1177729694. [DOI] [Google Scholar]
  • 7.Kalman R.E., Bucy R.S. New results in linear filtering and prediction theory. J. Basic Eng. 1961;83:95–108. doi: 10.1115/1.3658902. [DOI] [Google Scholar]
  • 8.DeMars K.J. Ph.D. Thesis. The University of Texas at Austin; Austin, TX, USA: 2010. Nonlinear Orbit Uncertainty Prediction and Rectification for Space Situational Awareness. [Google Scholar]
  • 9.DeMars K.J., Bishop R.H., Jah M.K. Entropy-based approach for uncertainty propagation of nonlinear dynamical systems. J. Guid. Control. Dyn. 2013;36:1047–1057. doi: 10.2514/1.58987. [DOI] [Google Scholar]
  • 10.Kim H., Liu B., Goh C.Y., Lee S., Myung H. Robust vehicle localization using entropy-weighted particle filter-based data fusion of vertical and road intensity information for a large scale urban area. IEEE Robot. Autom. Lett. 2017;2:1518–1524. doi: 10.1109/LRA.2017.2673868. [DOI] [Google Scholar]
  • 11.Zhang J., Du L., Ren M., Hou G. Minimum error entropy filter for fault detection of networked control systems. Entropy. 2012;14:505–516. doi: 10.3390/e14030505. [DOI] [Google Scholar]
  • 12.Liu Y., Wang H., Hou C. UKF based nonlinear filtering using minimum entropy criterion. IEEE Trans. Signal Process. 2013;61:4988–4999. doi: 10.1109/TSP.2013.2274956. [DOI] [Google Scholar]
  • 13.Julier S., Uhlmann J., Durrant-Whyte H.F. A new method for the nonlinear transformation of means and covariances in filters and estimators. IEEE Trans. Autom. Control. 2000;45:477–482. doi: 10.1109/9.847726. [DOI] [Google Scholar]
  • 14.Contreras-Reyes J.E., Cortés D.D. Bounds on rényi and shannon entropies for finite mixtures of multivariate skew-normal distributions: Application to swordfish (xiphias gladius linnaeus) Entropy. 2016;18:382. doi: 10.3390/e18110382. [DOI] [Google Scholar]
  • 15.Ren M., Zhang J., Fang F., Hou G., Xu J. Improved minimum entropy filtering for continuous nonlinear non-Gaussian systems using a generalized density evolution equation. Entropy. 2013;15:2510–2523. doi: 10.3390/e15072510. [DOI] [Google Scholar]
  • 16.Zhang Q. Performance enhanced Kalman filter design for non-Gaussian stochastic systems with data-based minimum entropy optimisation. AIMS Electron. Electr. Eng. 2019;3:382. doi: 10.3934/ElectrEng.2019.4.382. [DOI] [Google Scholar]
  • 17.Chen B., Dang L., Gu Y., Zheng N., Príncipe J.C. Minimum error entropy Kalman filter. IEEE Trans. Syst. Man Cybern. Syst. 2019 doi: 10.1109/TSMC.2019.2957269. [DOI] [Google Scholar]
  • 18.Gultekin S., Paisley J. Nonlinear Kalman filtering with divergence minimization. IEEE Trans. Signal Process. 2017;65:6319–6331. doi: 10.1109/TSP.2017.2752729. [DOI] [Google Scholar]
  • 19.Darling J.E., DeMars K.J. Minimization of the kullback–leibler divergence for nonlinear estimation. J. Guid. Control Dyn. 2017;40:1739–1748. doi: 10.2514/1.G002282. [DOI] [Google Scholar]
  • 20.Morelande M.R., Garcia-Fernandez A.F. Analysis of Kalman filter approximations for nonlinear measurements. IEEE Trans. Signal Process. 2013;61:5477–5484. doi: 10.1109/TSP.2013.2279367. [DOI] [Google Scholar]
  • 21.Raitoharju M., García-Fernández Á.F., Piché R. Kullback–Leibler divergence approach to partitioned update Kalman filter. Signal Process. 2017;130:289–298. doi: 10.1016/j.sigpro.2016.07.007. [DOI] [Google Scholar]
  • 22.Hu E., Deng Z., Xu Q., Yin L., Liu W. Relative entropy-based Kalman filter for seamless indoor/outdoor multi-source fusion positioning with INS/TC-OFDM/GNSS. Clust. Comput. 2019;22:8351–8361. doi: 10.1007/s10586-018-1803-1. [DOI] [Google Scholar]
  • 23.Yu W., Peng J., Zhang X., Li S., Liu W. An adaptive unscented particle filter algorithm through relative entropy for mobile robot self-localization. Math. Probl. Eng. 2013 doi: 10.1155/2013/567373. [DOI] [Google Scholar]
  • 24.Arasaratnam I., Haykin S. Cubature kalman filters. IEEE Trans. Autom. Control. 2009;54:1254–1269. doi: 10.1109/TAC.2009.2019800. [DOI] [Google Scholar]
  • 25.Kiani M., Barzegar A., Pourtakdoust S.H. Entropy-based adaptive attitude estimation. Acta Astronaut. 2018;144:271–282. doi: 10.1016/j.actaastro.2017.12.044. [DOI] [Google Scholar]
  • 26.Giffin A., Urniezius R. The Kalman filter revisited using maximum relative entropy. Entropy. 2014;16:1047–1069. doi: 10.3390/e16021047. [DOI] [Google Scholar]
  • 27.Chen B., Liu X., Zhao H., Principe J.C. Maximum correntropy Kalman filter. Automatica. 2017;76:70–77. doi: 10.1016/j.automatica.2016.10.004. [DOI] [Google Scholar]
  • 28.Chen B., Xing L., Liang J., Zheng N., Principe J.C. Steady-state mean-square error analysis for adaptive filtering under the maximum correntropy criterion. IEEE Signal Process. Lett. 2014;21:880–884. [Google Scholar]
  • 29.Gongmin Y., Jun W. Lectures on Strapdown Inertial Navigation Algorithm and Integrated Navigation Principles. Northwestern Polytechnical University Press; Xi’an, China: 2019. [Google Scholar]
  • 30.Kumari L., Padma Raju K. Application of Extended Kalman filter for a Free Falling body towards Earth. IJACSA Ed. 2011;2:4. doi: 10.14569/IJACSA.2011.020420. [DOI] [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES