Skip to main content
Entropy logoLink to Entropy
. 2023 Oct 13;25(10):1445. doi: 10.3390/e25101445

Jeffreys Divergence and Generalized Fisher Information Measures on Fokker–Planck Space–Time Random Field

Jiaxing Zhang 1
Editor: Jean-Pierre Gazeau1
PMCID: PMC10606917  PMID: 37895566

Abstract

In this paper, we present the derivation of Jeffreys divergence, generalized Fisher divergence, and the corresponding De Bruijn identities for space–time random field. First, we establish the connection between Jeffreys divergence and generalized Fisher information of a single space–time random field with respect to time and space variables. Furthermore, we obtain the Jeffreys divergence between two space–time random fields obtained by different parameters under the same Fokker–Planck equations. Then, the identities between the partial derivatives of the Jeffreys divergence with respect to space–time variables and the generalized Fisher divergence are found, also known as the De Bruijn identities. Later, at the end of the paper, we present three examples of the Fokker–Planck equations on space–time random fields, identify their density functions, and derive the Jeffreys divergence, generalized Fisher information, generalized Fisher divergence, and their corresponding De Bruijn identities.

Keywords: space–time random field, Fokker–Planck equations, differential entropy, Jeffreys divergence, fisher information, De Bruijn identities

1. Introduction

Information entropy and Fisher information are quantities to measure random information, and entropy divergence is derived from information entropy to measure the difference between two probability distributions. Formally, we can construct straightforward definitions of entropy divergence and Fisher information for the case of a space–time random field found on classical definitions. The density function, in their definitions, can be obtained in many different ways. In this paper, the density function of a space–time random field is obtained by Fokker–Planck equations. The traditional Fokker–Planck equation is a partial differential equation that describes the probability density function of a random process [1]. It describes the density function’s time-varying change rule. However, the Fokker–Planck equations for random fields, especially for space–time random fields, do not yet possess a distinct form. The classical equation needs to be generalized because the variable varies from time to space–time.

In this paper, we mainly obtain the relation between Jeffreys divergence and generalized Fisher information measure for space–time random field generated by Fokker–Planck equations. Jeffreys divergence is a symmetric entropy divergence, which is generalized from Kullback–Leibler divergence (KL divergence). Jeffreys divergence is a measure in information theory and statistics that evaluates the variation between anticipated and real probability distributions. However, if there is no overlap between the two distributions, the outcome will be infinite, which is a limitation of this approach. To prevent infinite results, we examine how Jeffreys divergence relates to generalized Fisher information for a space–time random field with slight variations in space–time parameters.

Moreover, the classical De Bruijn identity describes the relationship between differential entropy and the Fisher information of the Gaussian channel [2], and it can be generalized to other cases [3,4,5,6,7]. With gratitude to their works and following their ideas, we obtain De Bruijn identities on Jeffreys divergence and generalized Fisher information of space–time random fields, whose density functions satisfy Fokker–Planck equations.

1.1. Space–Time Random Field

The random field was first studied by Kolmogorov [8,9,10], and it was gradually improved by Yaglom [11,12,13] in the middle of the last century. A random field with nN+ variables can be expressed as

X(t1,t2,,tn) (1)

where (t1,t2,,tn)Rn. We call (1) a generalized random field or a multiparameter stochastic process. In some practical applications, we often use the concept of space–time random field. The space–time random field on a d-dimensional space is expressed as

X(t,x) (2)

where (t,x)R+×Rd are the space–time variables. It has many applications in statistics, finance, signal processing, stochastic partial differential equations, and other fields [14,15,16,17,18,19,20,21,22,23,24,25,26,27].

1.2. Kramers–Moyal Expansion and Fokker–Planck Equation

In the literature of stochastic processes, Kramers–Moyal expansion refers to a Taylor series of the master equation, named after Kramers and Moyal [28,29]. The Kramers–Moyal expansion is an infinite order partial differential equation

tp(u,t)=n=1(1)nn!nunKn(u,t)p(u,t) (3)

where p(u,t) is the density function and

Kn(u,t)=RuunW(u|u,t)du (4)

is the n-order conditional moment. Here, W(u|u,t) is the transition probability rate. The Fokker–Planck equation is obtained by keeping only the first two terms of the Kramers–Moyal expansion. In statistical mechanics, the Fokker–Planck equation is usually used to describe the time evolution of the probability density function of the velocity of a particle under the influence of drag forces and random forces, as in the famous Brownian motion, and this equation is commonly employed for determining the density function of an Itô stochastic differential equation [1].

1.3. Differential Entropy and De Bruijn Identity

The entropy of a continuous distribution was proposed by Shannon in 1948, known as differential entropy [30]:

hX=Rdp(x)logp(x)dx (5)

where h(·) represents the differential entropy and p(·) is the probability density function of X. However, differential entropy is not easy to calculate and seldom exists. There are related studies on the entropy of stochastic processes and continuous systems [31,32,33,34]. If we consider a classical one-dimensional Gaussian channel model

Yt=X+tG (6)

where X is the input signal, G is standard Gaussian noise, t0 is the strength, and Yt is the output, we can obtain that the density of Yt satisfies the following Fokker–Planck equation:

tp(y,t)=122y2p(y,t) (7)

Furthermore, the differential entropy of Yt can be calculated, and then its derivative with respect to t can be obtained as

dhYt(t)dt=12FIYt(t) (8)

where

FIYt(t)=Rlogp(y,t)y2p(y,t)dy (9)

is the Fisher information of Yt. The Equation (8) here is the De Burijn identity. The de Bruijn identity connects the differential entropy h(·) and the Fisher information FI(·), which shows that they are different aspects of the concept of “information”.

1.4. Entropy Divergence

In information theory and statistics, an entropy divergence is a statistical distance generated from information entropy to measure the difference between two probability distributions. There are various divergences generated by information entropy, such as Kullback–Leibler divergence [35], Jeffreys divergence [36], Jensen-Shannon divergence [37], and Rényi divergence [38]. These measures are applied in a variety of fields such as finance, economics, biology, signal processing, pattern recognition, and machine learning [39,40,41,42,43,44,45,46,47,48,49]. In this paper, we mainly focus on the Jeffreys divergence of two distributions, formed as

JDP,Q=Rp(u)q(u)logp(u)q(u)dμ(u) (10)

where μ is a measure of u.

2. Notations, Definitions, and Propositions

2.1. Notations and Assumptions

In this paper, we use the subsequent notations and definitions

  • Given a probability space (Ω,F,P), two real valued space–time random fields are denoted as X(ω;t,x), Y(ω;s,y) or X(t,x), Y(s,y), where ωΩ and (t,x)(s,y)R+×Rd, dN+ are space–time variables.

  • The probability density functions of P and Q are denoted as p and q. With uR, p(u;t,x) is the density value at (t,x) of X and q(u;s,y) is the density value at (s,y) of Y.

  • Unless there are specific restrictions on the ranges of variables, suppose that our density functions p(u;t,x) and q(u,s,y) belongs to C2,1,1(R×R+×Rd,R). This means that p(u;t,x) and q(u;s,y) are partial differentiable twice with respect to u and once with respect to (t,x) or (s,y), respectively.

  • Vectors that differ only from the k-th coordinate of x=(x1,x2,,xk,,xd) are denoted x˜(k)=(x1,x2,,xk,,xd), where the k-th coordinates are xk and xk, k=1,2,,d.

2.2. Definitions

To obtain the generalized De Bruijn identities between Jeffreys divergence and Fisher divergence, we need to introduce some new definitions and propositions.

The primary and most important measure of information is the Kullback–Leibler divergence for random fields. Definition 1 is easily obtained as follows.

Definition 1.

The Kullback–Leibler divergence between two space–time random fields X(t,x) and Y(s,y), (t,x),(s,y)R+×Rd, with density functions p(u;t,x) and q(u;s,y), is defined as

KLP(t,x)Q(s,y)=Rp(u;t,x)logp(u;t,x)q(u;s,y)du (11)

Similar to the classical Kullback–Leibler divergence, Kullback–Leibler divergence on random fields is not symmetrical, i.e.,

KLP(t,x)Q(s,y)KLQ(s,y)P(t,x) (12)

Following the classical definition of Jeffreys divergence on two random variables, we mainly consider Jeffreys divergence for random fields in this paper.

Definition 2.

The Jeffreys divergence between space–time random fields X(t,x) and Y(s,y), (t,x),(s,y)R+×Rd, with density function p(u;t,x) and q(u;s,y) is defined as

JDP(t,x),Q(s,y)=KLP(t,x)Q(s,y)+KLQ(s,y)P(t,x) (13)

Here, we replace ∥ with, in the distortion measure to emphasize the symmetric property.

Another significant measure of information is Fisher information. In this paper, we consider the generalized Fisher information of the space–time random field.

Definition 3.

The Generalized Fisher information of the space–time random field X(t,x), (t,x)R+×Rd, with density function p(u;t,x) defined by nonnegative function f(·), is formed as

FIfP(t,x)=Rf(u)ulogp(u;t,x)2p(u;t,x)du (14)

In this case, where f is equal to 1, FI1(P(t,x)) represents the typical Fisher information. In addition to Equation (14), there are similar forms of generalized Fisher information

FIf(t)P(t,x)=Rf(u)tlogp(u;t,x)2p(u;t,x)du (15)

and

FIf(xk)P(t,x)=Rf(u)xklogp(u;t,x)2p(u;t,x)du (16)

for k=1,2,,d.

Obviously, (15) and (16) are generalized Fisher information on space–time variables. Regarding the generalized Fisher information (14), we can come to a following simple proposition.

Proposition 1.

For the arbitrary positive continuous function f(·), suppose the generalized Fisher information of continuous random variable X

FIf(X):=Rf(u)dlogpX(u)du2pX(u)du (17)

is well defined, where pX(u) represents the probability density. Then, we have the generalized Fisher information inequality

1FIf(X+Y)1FIf(X)+1FIf(Y) (18)

when f1, FI1(X) represents the Fisher information in the standard case.

Proof. 

Denote Z=X+Y, pX, pY, and pZ represent densities, i.e.,

pZ(z)=RpX(x)pY(zx)dx (19)

and derivative function

pZ(z)=RpX(x)pY(zx)dx (20)

If pX, pY, and pZ never vanish,

pZ(z)pZ(z)=RpX(x)pY(zx)pZ(z)dx=RpX(x)pY(zx)pZ(z)pX(x)pX(x)dx=EpX(x)pX(x)|Z=z (21)

is the conditional expectation of pX(x)pX(x) for given z. Similarly, we can obtain

pZ(z)pZ(z)=EpY(y)pY(y)|Z=z (22)

and μ, λR, we also find that

EμpX(x)pX(x)+λpY(y)pY(y)|Z=z=(μ+λ)pZ(z)pZ(z) (23)

Then, we have

(μ+λ)pZ(z)pZ(z)2=EμpX(x)pX(x)+λpY(y)pY(y)|Z=z2EμpX(x)pX(x)+ypY(y)pY(y)2|Z=z (24)

with equality only if

μpX(x)pX(x)+λpY(y)pY(y)=(μ+λ)pZ(z)pZ(z) (25)

with probability 1 whenever z=x+y and we have

f(z)(μ+λ)pZ(z)pZ(z)2f(z)EμpX(x)pX(x)+ypY(y)pY(y)2|Z=z (26)

Averaging both sides over the distribution of z

Ef(z)(μ+λ)pZ(z)pZ(z)2Ef(z)EμpX(x)pX(x)+ypY(y)pY(y)2|Z=z=μ2Ef(z)EpX(x)pX(x)2|Z=z+λ2EEf(z)pY(y)pY(y)2|Z=z (27)

i.e.,

(μ+λ)2FIf(X+Y)μ2FIf(X)+λ2FIf(Y) (28)

Let μ=1FIf(X) and λ=1FIf(Y), we obtain

1FIf(X+Y)1FIf(X)+1FIf(Y) (29)

According to Definition 3, we can obtain relevant definitions on the generalized Fisher information measure.

Definition 4.

The generalized Cross–Fisher information for space–time random fields X(t,x) and Y(s,y), (t,x),(s,y)R+×Rd, with density functions p(u;t,x) and q(u;s,y), defined by the nonnegative function f(·), is defined as

CFIf(P(t,x),Q(s,y))=Rf(u)ulogq(u;s,y)2p(u;t,x)du (30)

Similar to the concept of cross-entropy, it is easy to verify that (30) is symmetrical about P and Q.

Definition 5.

The generalized Fisher divergence for space–time random fields X(t,x) and Y(s,y), for (t,x),(s,y)R+×Rd, with density functions p(u;t,x) and q(u;s,y), defined by nonnegative function f(·), is defined as

FDfP(t,x)Q(s,y)=Rf(u)ulogp(u;t,x)ulogq(u;s,y)2p(u;t,x)du (31)

In particular, when f1, FD1P(t,x)Q(s,y) represents the typical Fisher divergence.

Obviously, the generalized Fisher divergence between two random fields is not a symmetrical measure of information. We need to create a new formula to expand on (31) in order to achieve symmetry.

Definition 6.

The generalized Fisher divergence for space–time random fields X(t,x) and Y(s,y), (t,x),(s,y)R+×Rd, with density functions p(u;t,x) and q(u;s,y), defined by nonnegative functions f(·) and g(·), is defined as

FD(f,g)P(t,x)Q(s,y)=Rf(u;t,x)ulogp(u;t,x)g(u;s,y)ulogq(u;s,y)×ulogp(u;t,x)ulogq(u;s,y)p(u;t,x)+q(u,s,y)du (32)

In particular, if f equals g, the generalized Fisher divergence for random fields using a single function is denoted as FD(f,f)P(t,x)Q(s,y).

In general, FD(f,g)P(t,x)Q(s,y) is asymmetric with respect to P and Q, i.e.,

FD(f,g)P(t,x)Q(s,y)FD(f,g)Q(s,y)P(t,x) (33)

If we suppose that f and g are functions only related to P and Q, i.e.,

f(u;t,x)=Tp(t,x)(u)g(u;s,y)=Tq(s,y)(u) (34)

where T is an operator; the generalized Fisher divergence FD(f,g)P(t,x)Q(s,y) can be rewritten as

FD(f,g)P(t,x)Q(s,y)=RTp(t,x)(u)ulogp(u;t,x)Tq(s,y)(u)ulogq(u;s,y)×ulogp(u;t,x)ulogq(u;s,y)p(u;t,x)+q(u,s,y)du (35)

and we can easily obtain

FD(f,g)P(t,x)Q(s,y)=FD(g,f)Q(t,x)P(s,y) (36)

In this case, we call (35) symmetric Fisher divergence for random fields generated by operator T and denote it as

sFDTP(t,x),Q(s,y)=RTp(t,x)(u)ulogp(u;t,x)Tq(s,y)(u)ulogq(u;s,y)×ulogp(u;t,x)ulogq(u;s,y)p(u;t,x)+q(u,s,y)du (37)

Notice that

AaBb=12×2AaBb=12AaAb+AbBb+AaBa+BaBb=12A+Bab+ABa+b (38)

for A, B, a, bR; then, we can rewrite (37) as

sFDTP(t,x),Q(s,y)=12RTp(t,x)(u)+Tq(s,y)(u)ulogp(u;t,x)ulogq(u;s,y)2×p(u;t,x)+q(u,s,y)du+12RTp(t,x)(u)Tq(s,y)(u)ulogp(u;t,x)2ulogq(u;s,y)2×p(u;t,x)+q(u,s,y)du=12FDTp(t,x)+Tq(s,y)P(t,x)Q(s,y)+FDTp(t,x)+Tq(s,y)Q(s,y)P(t,x)+12FITp(t,x)Tq(s,y)P(t,x)+FITp(t,x)Tq(s,y)Q(s,y)+12CFITp(t,x)Tq(s,y)Q(s,y),P(t,x)CFITp(t,x)Tq(s,y)P(t,x),Q(s,y) (39)

Lemma 1

(Kramers–Moyal expansion [28,29]). Suppose that the random process X(t) has any order moment; then, the probability density function p(u,t) satisfies the Kramers–Moyal expansion

tp(u,t)=n=1(1)nn!nunKn(u,t)p(u,t) (40)

where

Kn(u,t)=RuunW(u|u,t)du (41)

is the n-order conditional moment and W(u|u,t) is the transition probability rate.

Lemma 2

(Pawula theorem [50,51]).If the limit on conditional moment of random process X(t)

limΔt01ΔtEX(t+Δt)X(t)n|X(t)=x (42)

exists for all nN+, and the limit value equals 0 for some even number, then the limit values are 0 for all n3.

The Pawula theorem states that there are only three possible cases in the Kramers–Moyal expansion:

  • (1)

    The Kramers–Moyal expansion is truncated at n=1, meaning that the process is deterministic;

  • (2)

    The Kramers–Moyal expansion stops at n=2, with the resulting equation being the Fokker–Planck equation, and describes diffusion processes;

  • (3)

    The Kramers–Moyal expansion contains all the terms up to n=.

In this paper, we only focus on the case of the Fokker–Planck equation.

3. Main Results and Proofs

In this section, we establish the Fokker–Planck equations for continuous space–time random field. Additionally, we present the relationship theorem between Jeffreys divergence and Fisher information, as well as the De Bruijn identities connection between Jeffreys divergence and Fisher divergence.

Theorem 1.

The probability density function p(u;t,x) of the continuous space–time random field X(t,x), uR, (t,x)R+×Rd satisfies the following Fokker–Planck equations:

tp(u;t,x)=122u2b0(u;t,x)p(u,t,x)ua0(u;t,x)p(u;t,x)xkp(u;t,x)=122u2bk(u;t,x)p(u,t,x)uak(u;t,x)p(u;t,x)k=1,2,,d (43)

where

a0(u;t,x)=limΔt01ΔtM1(u;t,Δt,x)b0(u;t,x)=limΔt01ΔtM2(u;t,Δt,x)ak(u;t,x)=limΔxk01ΔxkM˜1(u;t,x,Δxk)bk(u;t,x)=limΔxk01ΔxkM˜2(u;t,x,Δxk)k=1,2,,d (44)

here,

Mn(u;t,Δt,x)=EX(t+Δt,x)X(t,x)n|X(t,x)=uM˜n(u;t,x,Δxk)=EX(t,x+Δxkek)X(t,x)n|X(t,x)=u (45)

are n-order conditional moments and ek=(0,0,,1,,0)Rd are standard orthogonal basis vectors, k=1,2,,d.

Proof. 

Δt0, we can obtain the difference of density function in the time variable

p(u;t+Δt,x)p(u;t,x)=n=1+(1)nn!nunMn(u;t,Δt,x)p(u;t,x) (46)

where

Mn(u;t,Δt,x)=EX(t+Δt,x)X(t,x)n|X(t,x)=u (47)

is the n-order conditional moment. Then, the partial derivative of the density function with respect to t is

tp(u;t,x)=limΔt01Δtn=1+(1)nn!nunMn(u;t,Δt,x)p(u;t,x) (48)

The Pawula theorem implies that if the Kramers–Moyal expansion stops after the second term, we obtain the Fokker–Planck equation about the time variable t

tp(u;t,x)=122u2b0(u;t,x)p(u,t,x)ua0(u;t,x)p(u;t,x) (49)

where

a0(u;t,x)=limΔt01ΔtM1(u;t,Δt,x)b0(u;t,x)=limΔt01ΔtM2(u;t,Δt,x) (50)

Similarly, we may consider the increment Δxk of the spatial variable xk, and we can obtain the Fokker–Planck equations about xk as

xkp(u;t,x)=122u2bk(u;t,x)p(u,t,x)uak(u;t,x)p(u;t,x) (51)

where

ak(u;t,x)=limΔxk01ΔxkM˜1(u;t,x,Δxk)bk(u;t,x)=limΔxk01ΔxkM˜2(u;t,x,Δxk) (52)

here,

M˜n(u;t,x,Δxk)=EX(t,x+Δxkek)X(t,x)n|X(t,x)=u (53)

ek=(0,0,,1,,0)Rd are standard orthogonal basis vectors, k=1,2,,d. □

The Fokker–Planck equations are partial differential equations that describe the probability density function of the space–time random field, similar to the classical Fokker–Planck equation. Solving a system of partial differential equations for general Fokker–Planck equations proves to be challenging. Fortunately, in Section 4 we present three distinct categories of space–time random fields in detail, along with their corresponding Fokker–Planck equations, and deduce their probability density functions.

Next, we examine the relationship between Jeffreys divergence and Fisher information in a single space–time random field when there are different time or spatial variables.

Theorem 2.

Suppose that p(u;t,x)>0 is a continuous differential density function of the space–time random field X(t,x), the partial derivatives up(u;t,x), tp(u;t,x), xkp(u;t,x) are continuous bounded functions, and the integrals in the proof are well-defined, k=1,2,,d, uR, (t,x)R+×Rd. Then, we have

limts0JDP(t,x),P(s,x)|ts|2=FI1tX(t,x)limxkxk0JDP(t,x),P(t,x˜(k))|xkxk|2=FI1xkX(t,x)k=1,2,,d (54)

Proof. 

For fixed xR, and st>0,

JDP(t,x),P(s,x)=KLP(t,x)||P(s,x)+KLP(s,x)||P(t,x)=Rlogp(u;t,x)logp(u,s,x)p(u;t,x)p(u,s,x)du (55)

then we can obtain

limts0JDP(t,x),P(s,x)|ts|2=limts0Rlogp(u;t,x)logp(u,s,x)tsp(u;t,x)p(u,s,x)tsdu (56)

Notice that

limts0logp(u;t,x)logp(u,s,x)ts=tlogp(u;t,x)limts0p(u;t,x)p(u,s,x)ts=tp(u;t,x) (57)

exist, and we obtain

limts0JDP(t,x),P(s,x)|ts|2=Rtlogp(u;t,x)tp(u;t,x)du=Rtlogp(u;t,x)2p(u;t,x)du=FI1tX(t,x) (58)

Similarly, for fixed t and xkxk, we can obtain the identity on Jeffreys divergence and Fisher information for space coordinates

limxkxk0JDP(t,x),P(t,x˜(k))|xkxk|2=FI1xkX(t,x) (59)

for k=1,2,,d. □

Theorem 2 states that as the space–time variable difference approaches zero, the Fisher information of the space–time random field is the limit of the ratio of Jeffreys divergence at different locations to the square of space–time variable difference. It is noteworthy that Theorem 2 specifically addresses Jeffreys divergence only in cases where a single space–time random field is situated in distinct space–time positions, and where the difference between space–time variables approaches to 0. This ensures that Jeffreys divergence will not be infinite.

Theorem 3.

Suppose that p(u;t,x) and q(u;t,x) are continuous differentiable density functions of space–time random fields X(t,x) and Y(t,x) such that

limu12ubk(1)(u;t,x)p(u;t,x)ak(1)(u;t,x)p(u;t,x)logp(u;t,x)q(u;t,x)q(u;t,x)p(u;t,x)=0limu12ubk(2)(u;t,x)q(u;t,x)ak(2)(u;t,x)q(u;t,x)logq(u;t,x)p(u;t,x)p(u;t,x)q(u;t,x)=0 (60)

where ak, bk are the forms in (44) and (45), and (t,x)R+×Rd, k=0,1,2,,d. Then, the Jeffreys divergence JDP(t,x),Q(t,x) satisfies generalized De Bruijn identities

tJDP(t,x),Q(t,x)=12FDb0(1),b0(2)P(t,x)Q(t,x)R0P(t,x)Q(t,x)xkJDP(t,x),Q(t,x)=12FDbk(1),bk(2)P(t,x)Q(t,x)RkP(t,x)Q(t,x)k=1,2,,d (61)

where

R0P(t,x)Q(t,x)=R12uu2b0(1)b0(2)ua0(1)a0(2)p+qduRkP(t,x)Q(t,x)=R12uu2bk(1)bk(2)uak(1)ak(2)p+qduk=1,2,,d (62)

here, we omit (u;t,x) in the integrals for convenience.

Proof. 

By Definition 2, we have

JDP(t,x),Q(t,x)=KL(P(t,x)Q(t,x))+KL(Q(t,x)P(t,x))=Rplogpqdu+Rqlogqpdu=Rplogpq+qlogqpdu (63)

where p:=p(u;t,x), q:=q(u;t,x) are density functions of X(t,x) and Y(t,x); here, we omit (u;t,x).

Notice that

upq=1quppququqp=1puqqpup (64)

i.e.,

pquq=upqupqqpup=uqpuqp (65)

and

ulogpulogqp+q=1pup1quqp+q=upuq+qpuppquq=pquppuqp2+qquppuqq2=puqpqupq (66)

then,

tJDP(t,x),Q(t,x)=Rtplogpq+qtpq+tqlogqp+ptqpdu=Rtplogpq+tppqtq+tqlogqp+tqqptpdu=Rlogpqqptp+logqppqtqdu=Rlogpqqp12uu2b0(1)pua0(1)pdu+Rlogqppq12uu2b0(2)qua0(2)qdu=R12ub0(1)pa0(1)pqpupquqpduR12ub0(2)qa0(2)qpquqpupqdu=R12ub0(1)pa0(1)p1pqupqpuqpduR12ub0(2)qa0(2)q1qpuqpqupqdu=R12ub0(1)pa0(1)p1p12ub0(2)qa0(2)q1q×ulogpulogqp+qdu=R12b0(1)ulogp12b0(2)ulogq+12ub0(1)b0(2)a0(1)a0(2)×ulogpulogqp+qdu=R12b0(1)ulogp12b0(2)ulogqulogpulogqp+qduR12ub0(1)b0(2)a0(1)a0(2)ulogpulogqp+qdu=12FDb0(1),b0(2)P(t,x)Q(t,x)R0P(t,x)Q(t,x) (67)

where

FDb0(1),b0(2)P(t,x)Q(t,x)=Rb0(1)ulogpb0(2)ulogqulogpulogqp+qdu (68)

and

R0P(t,x)Q(t,x)=R12uu2b0(1)b0(2)ua0(1)a0(2)p+qdu (69)

Similarly, for k=1,2,,d, we can obtain the generalized De Bruijn identities about the spatial variable xk

xkJDP(t,x),Q(t,x)=12FDbk(1),bk(2)P(t,x)Q(t,x)Rk(P(t,x),Q(t,x)) (70)

where

RkP(t,x)Q(t,x)=R12uu2bk(1)bk(2)uak(1)ak(2)p+qdu (71)

then we obtain the conclusion. □

Unlike Theorem 3, Theorem 4 focuses on the Jeffreys divergence between two separate space–time random fields X(t,x) and Y(t,x), both at the same position (t,x), and establishes the identities of the connection between the Jeffreys divergence and the Fisher divergence of X(t,x) and Y(t,x). This is known as the De Bruijn identities. To prevent Jeffreys divergence from becoming infinite, it is necessary for the difference between the probability density functions of X(t,x) and Y(t,x) to be small. In Section 4, we obtain Jeffreys divergence and Fisher divergence using the same type of Fokker–Planck equations but with different parameters. This allows for the selection of only the appropriate parameters.

4. Three Fokker–Planck Random Fields and Their Corresponding Information Measures

In this section, we present three types of Fokker–Planck equations and derive their corresponding density functions and information measures, which are Jeffreys divergence, generalized Fisher information, and Fisher divergence. With these quantities, the results corresponding to the applications of Theorems 2 and 3 are obtained. On the one hand, we calculate the ratio of Jeffreys divergence to the square of space–time variation on the identical Fokker–Planck space–time random field at various space–time points, in comparison to generalized Fisher information. On the other hand, we derive the De Burijn identities for Jeffreys divergence and generalized Fisher divergence from Fokker–Planck equations on a single space–time random field at the corresponding space–time location, under same type but with different parameters.

First, we present a theorem regarding simple type Fokker–Planck equations of the random field.

Theorem 4.

Suppose the functions in the Fokker–Planck Equations (43) for the continuous random field X(t,x) are formulated as follows:

a0(u;t,x)=a0(t,x)b0(u;t,x)=b0(t,x)>0ak(u;t,x)=ak(t,x)bk(u;t,x)=bk(t,x)>0k=1,2,,d (72)

where a0, ak, b0, and bk are continuously differentiable functions independent of u and two continuously differentiable functions, α(t,x) and β(t,x), exist such that

dα(t,x)=a0dt+a1dx1++addxddβ(t,x)=b0dt+b1dx1++bddxd (73)

the initial density function is p(u;t,x)=δuu0(x) as prod(t,x)=0; then, the density function of X(t,x) is presented as follows:

p(u;t,x)=12πβ(t,x)euu0(x)α(t,x)22β(t,x) (74)

Proof. 

It can be easily inferred that the Fokker–Planck equations are simple parabolic equations, and their solution can be obtained through Fourier transform

p(u;t,x)=12π0tb0(s,x)dseuu0(x)0ta0(s,x)ds220tb0(s,x)dsp(u;t,x)=12π0xkbk(t,x)dxkeuu0(x)0xkak(t,x)dxk220xkbk(t,x)dxk (75)

Recall that there are two functions α(t,x) and β(t,x) such that

dα(t,x)=a0(t,x)dt+a1(t,x)dx1++ad(t,x)dxddβ(t,x)=b0(t,x)dt+b1(t,x)dx1++bd(t,x)dxd (76)

we can obtain the probability density function

p(u;t,x)=12πβ(t,x)euu0(x)α(t,x)22β(t,x) (77)

Actually, numerous examples exist in which the Fokker–Planck equations comply with Theorem 4. Let B(t,x) be the (1+d,1) Brownian sheet [52,53], that is, a centered continuous Gaussian process that is indexed by (1+d) real, positive parameters and takes its values in R. Moreover, its covariance structure is given by

EB(t,x)B(s,y)=tsk=1dxkyk (78)

for (t,x1,x2,,xd), (s,y1,y2,,yd)R+×R+d, where ·· represents the minimum of two numbers. We can easily obtain

EB2(t,x)=prod(t,x) (79)

where prod(t,x)=tx1x2xd is the coordinate product of (t,x) and the density function is

p(1)(u;t,x)=12πprod(t,x)eu22prod(t,x) (80)

Moreover, the Fokker–Planck equations of Brownian sheet are

tp(1)(u;t,x)=prod(x)22u2p(1)(u,t,x)xkp(1)(u;t,x)=prod(t,x)2xk2u2p(1)(u,t,x)k=1,2,,d (81)

with the initial condition p(u;t,x)=δ(u) as prod(t,x)=0.

Following the concept of constructing a Brownian bridge on Brownian motion [53], we refer to

B(t,x)=B(t,x)prod(t,x)B(1,1,,1) (82)

as a Brownian sheet bridge on the cube (t,x)[0,1]×[0,1]d, where B(t,x) represents the Brownian sheet. Obviously, B(t,x) is Gaussian, and EB(t,x)=0 and the covariance structure are

EB(t,x)B(s,y)=EB(t,x)B(s,y)prod(t,x)prod(s,y) (83)

we can obtain

EB2(t,x)=prod(t,x)1prod(t,x) (84)

and the density function of B(t,x) is

p(2)(u;t,x)=12πprod(t,x)1prod(t,x)eu22prod(t,x)1prod(t,x) (85)

In addition to this, the Fokker–Planck equations of Brownian sheet bridge are

tp(2)(u;t,x)=prod(x)12prod(t,x)22u2p(2)(u,t,x)xkp(2)(u;t,x)=prod(t,x)2xk12prod(t,x)2u2p(2)(u,t,x)k=1,2,,d (86)

with the initial condition p(u;t,x)=δ(u) as prod(t,x)=0, and we obtain the solution (85).

Combining two probability density functions (80) and (85) yields their respective Jeffreys divergence and generalized De Burijn identities. The Jeffreys divergence of (74) can be obtained at various space–time points as

JDP(t,x),P(s,y)=α(t,x)α(s,y)2+β(s,y)2β(t,x)+α(t,x)α(s,y)2+β(t,x)2β(s,y)1 (87)

and the Fisher divergence between P(1) and P(2) at the identical space–time point represents

FDbk(1),bk(2)P(1)(t,x)P(2)(t,x)=1β12(t,x)β22(t,x){α1(t,x)α2(t,x)2bk(2)β12(t,x)+bk(1)β22(t,x)+β1(t,x)β2(t,x)β1(t,x)+β2(t,x)bk(2)β1(t,x)bk(1)β2(t,x)} (88)

where k=0,1,,d.

Bring the density function of Brownian sheet into Equation (87); we can easily obtain the Jeffreys divergence of the Brownian sheet at various space–time points as

JDP(1)(t,x),P(1)(s,y)=prod(s,y)2prod(t,x)+prod(t,x)2prod(s,y)1 (89)

and the generalized Fisher information on space–time variables is as follows:

FI1(t)P(1)(t,x)=12t2FI1(xk)P(1)(t,x)=12xk2 (90)

k=1,2,,d.

Then, we can obtain quotients of the squared difference between Jeffreys divergence and space–time variables

JDP(1)(t,x),P(1)(s,x)|ts|2=12stJDP(1)(t,x),P(1)(t,x˜(k))|xkxk|2=12xkxk (91)

and then we can obtain the relation between quotients and generalized Fisher information

JDP(1)(t,x),P(1)(s,x)|ts|2FI1(t)P(1)(t,x)=tsJDP(1)(t,x),P(1)(t,x˜(k))|xkxk|2FI1(xk)P(1)(t,x)=xkxk (92)

for k=1,2,,d. If we consider the approximation of spacetime points (t,x) and (s,y), the final result (92) satisfies the conclusion of Theorem 2.

Similarly, we can obtain the Jeffreys divergence of Brownian sheet bridge at different space–time points

JDP(2)(t,x),P(2)(s,y)=prod(s,y)1prod(s,y)2prod(t,x)1prod(t,x)+prod(t,x)1prod(t,x)2prod(s,y)1prod(s,y)1 (93)

and the generalized Fisher information on space–time variables

FI1(t)P(2)(t,x)=12prod(t,x)22t21prod(t,x)2FI1(xk)P(2)(t,x)=12prod(t,x)22xk21prod(t,x)2 (94)

k=1,2,,d. Further, we can easily get the quotients of the squared difference between Jeffreys divergence and space–time variables

JDP(2)(t,x),P(2)(s,x)|ts|2=1prod(x)(s+t)22st1prod(s,x)1prod(t,x)JDP(2)(t,x),P(2)(t,x˜(k))|xkxk|2=12xkxk1prod(t,x)1prod(t,x˜(k))1prod(t,x)xkxk+xk2 (95)

and then we can obtain the relation between quotients and generalized Fisher information

JDP(2)(t,x),P(2)(s,x)|ts|2FI1(t)P(2)(t,x)=t1prod(t,x)1prod(x)(s+t)2s1prod(s,x)12prod(t,x)2JDP(2)(t,x),P(2)(t,x˜(k))|xkxk|2FI1(xk)P(2)(t,x)=xk1prod(t,x)xk1prod(t,x˜(k))12prod(t,x)21prod(t,x)xkxk+xk2 (96)

for k=1,2,,d. Without loss of generality, the result (96) also satisfies Theorem 2.

Next, we evaluate the Jeffreys divergence between the density functions (80) and (85) for the same space–time points. It should be noted that the Brownian sheet bridge density function is defined on a bounded domain; therefore, we limit our analysis to the space–time region of (t,x)[0,1]×[0,1]d.

The Jeffreys divergence between P(1) and P(2) can be easily obtained as

JDP(1)(t,x),P(2)(t,x)=1prod(t,x)2+121prod(t,x)1 (97)

and the Fisher divergence as shown in (88) is given by

FDb0(1),b0(2)P(1)(t,x)P(2)(t,x)=prod(x)prod(x)1prod(t,x)2FDbk(1),bk(2)P(1)(t,x)P(2)(t,x)=prod(t,x)xkprod(t,x)xk1prod(t,x)2 (98)

with the remainder terms

R0P(1)(t,x)P(2)(t,x)=RkP(1)(t,x)P(2)(t,x)=0 (99)

for k=1,2,,d. Furthermore, we can obtain the generalized De Bruijn identities

tJDP(1)(t,x),P(2)(t,x)=12FDb0(1),b0(2)P(1)(t,x)P(2)(t,x)xkJDP(1)(t,x),P(2)(t,x)=12FDbk(1),bk(2)P(1)(t,x)P(2)(t,x)k=1,2,,d (100)

Next, we present two categories of significant Fokker–Planck equations and provide pertinent illustrations for computing Jefferys divergence, Fisher information, and Fisher divergence.

Theorem 5.

Suppose the functions in the Fokker–Planck Equations (43) for the continuous random field X(t,x) are formulated as follows:

ak(u;t,x)0bk(u;t,x)=bk(t,x)u2>0k=0,1,2,,d (101)

where bk are continuously differentiable functions independent of u and a continuously differentiable function β(t,x) exists, such that

dβ(t,x)=b0(t,x)dt+b1(t,x)dx1++bd(t,x)dxd (102)

the initial value X(t,x)=1 as prod(t,x)=0 and the initial density function is p(u;t,x)=δ(u1) as prod(t,x)=0. Then, the density function is

p(u;t,x)=eβ(t,x)2πβ(t,x)elogu+32β(t,x)22β(t,x) (103)

Proof. 

Depending on the conditions, it is easy to obtain the Fokker–Planck equations as

tp(u;t,x)=b0(t,x)2u22p(u;t,x)u2+2b0(t,x)uup(u;t,x)+b0(t,x)p(u;t,x)xkp(u;t,x)=bk(t,x)2u22p(u;t,x)u2+2bk(t,x)uup(u;t,x)+bk(t,x)p(u;t,x)k=1,2,,d (104)

Take the transformation v=logu or u=ev and note p˜(v;t,x)=p(u(v);t,x); we can obtain

tp˜(v;t,x)=b0(t,x)22v2p˜(v;t,x)+3b0(t,x)2vp˜(v;t,x)+b0(t,x)p˜(v;t,x)xkp˜(v;t,x)=bk(t,x)22v2p˜(v;t,x)+3bk(t,x)2vp˜(v;t,x)+bk(t,x)p˜(v;t,x)k=1,2,,d (105)

with the solution

p˜(v;t,x)=e0tb0(s,x)ds2π0tb0(s,x)dsev+320tb0(s,x)ds220tb0(s,x)dsp˜(v;t,x)=e0xkbk(t,x)dxk2π0xkbk(t,x)dxkev+320xkbk(t,x)dxk220xkbk(t,x)dxkk=1,2,,d (106)

then,

p(u;t,x)=e0tb0(s,x)ds2π0tb0(s,x)dselogu+320tb0(s,x)ds220tb0(s,x)dsp(u;t,x)=e0xkbk(t,x)dxk2π0xkbk(t,x)dxkelogu+320xkbk(t,x)dxk220xkbk(t,x)dxkk=1,2,,d (107)

Recall that a continuously differential function β(t,x) exists such that

dβ(t,x)=b0(t,x)dt+b1(t,x)dx1++bd(t,x)dxd (108)

this enables the derivation of the probability density

p(u;t,x)=eβ(t,x)2πβ(t,x)elogu+32β(t,x)22β(t,x) (109)

Remark 1.

In the stochastic process theory, a correlation exists between the Fokker–Planck equation and the Itô process. Specifically, if the Itô process is

dXt=μ(Xt,t)dt+σ(Xt,t)dBt (110)

then the corresponding Fokker–Planck equation can be obtained as

tp(u,t)=122u2σ2(u,t)p(u,t)uμ(u,t)p(u,t) (111)

where μ and σ represent drift and diffusion, Bt is the standard Brownian motion, or

dXtdt=μ(Xt,t)+σ(Xt,t)Wt (112)

where Wt=dBtdt represents the white noise. Actually, if we consider the Itô processes corresponding to Fokker–Planck equations from Theorem 5, we can obtain

tX(t,x)=b0(t,x)X(t,x)WtxkX(t,x)=bk(t,x)X(t,x)Wk (113)

where Wk represents the space white noise with respect to xk, k=1,2,,d. Further, we can also write Equation (113) in vector form

X(t,x)=γ(t,x)X(t,x)W(t,x) (114)

where

γ(t,x)=b0(t,x),b1(t,x),,bd(t,x)W(t,x)=Wt,W1,,Wd (115)

represents the gradient operator andrepresents element by element multiplication. Notice that each equation in Equation (113) is similar in form to the geometric Brownian motion in the theory of stochastic processes. Similarly, we can call the space–time random field that satisfies Equation (113) a geometric Brownian filed.

If we consider different β3(t,x) and β4(t,x) in density function (103), we can obtain density functions p(3)(u;t,x) and p(4)(u;t,x); then, we can obtain the Jeffreys divergence

JDP(3)(t,x),P(3)(s,y)=β3(t,x)+β3(s,t)+48β3(t,x)β3(s,y)β3(t,x)β3(s,y)2 (116)

and generalized Fisher information

FI1(t)P(3)(t,x)=β3(t,x)+24β32(t,x)b0(3)(t,x)2FI1(xk)P(3)(t,x)=β3(t,x)+24β32(t,x)bk(3)(t,x)2 (117)

and then the quotients

JDP(3)(t,x),P(3)(s,x)|ts|2=β3(t,x)+β3(s,x)+48β3(t,x)β3(s,y)β3(t,x)β3(s,y)ts2JDP(3)(t,x),P(3)(t,x˜(k))|xkxk|2=β3(t,x)+β3(t,x˜(k))+48β3(t,x)β3(t,x˜(k))β3(t,x)β3(t,x˜(k))xkxk2 (118)

and we can easily obtain the relation

JDP(3)(t,x),P(3)(s,y)|ts|2FI1(t)P(3)(t,x)=β3(t,x)β3(s,x)β3(t,x)+β3(s,x)+42β3(t,x)+2β3(t,x)β3(s,y)b0(3)(t,x)ts2JDP(3)(t,x),P(3)(t,x˜(k))|xkxk|2FI1(xk)P(3)(t,x)=β3(t,x)β3(t,x˜(k))β3(t,x)+β3(t,x˜(k))+42β3(t,x)+2β3(t,x)β3(t,x˜(k))bk(3)(t,x)xkxk2 (119)

for k=1,2,,d. Without a loss of generality, the result (119) also corroborates Theorem 2.

Furthermore, if we consider different β3(t,x) and β4(t,x) in density function (103), we can obtain density functions p(3)(u;t,x) and p(4)(u;t,x); then, the generalized Fisher divergence at the same space–time points is

FDbk(3),bk(4)P(3)(t,x)P(4)(t,x)=β3(t,x)β4(t,x)4β32(t,x)β42(s,y)×bk(4)(t,x)β3(t,x)β32(t,x)2β42(s,y)bk(3)(t,x)β4(t,x)β42(s,y)2β32(t,x)×4β3(t,x)+4β4(t,x)3β3(t,x)β4(t,x)bk(4)(t,x)β3(t,x)bk(3)(t,x)β4(t,x) (120)

with the remainder terms

RkP(3)(t,x)P(4)(t,x)=2bk(3)(t,x)bk(4)(t,x) (121)

k=0,1,2,,d. Then, the generalized De Bruijn identities are as follows:

tJDP(3)(t,x),P(4)(t,x)=12FDb0(3),b0(4)P(3)(t,x)P(4)(t,x)2b0(3)(t,x)b0(4)(t,x)xkJDP(3)(t,x),P(4)(t,x)=12FDbk(3),bk(4)P(3)(t,x)P(4)(t,x)2bk(3)(t,x)bk(4)(t,x)k=1,2,,d (122)

Additionally, we offer an alternative non-trivial form below that utilizes the implicit functions method to express our results. This form differs from the one presented in Theorem 5.

Theorem 6.

Suppose the functions in Fokker–Planck Equations (43) for the continuously bounded random field X(t,x)[0,1] are formulated as follows:

ak(u;t,x)=32bk(t,x)ubk(u;t,x)=bk(t,x)1u2k=0,1,2,,d (123)

where bk are continuously differentiable functions independent of u and a continuously differentiable function β(t,x) exists such that

dβ(t,x)=b0(t,x)dt+b1(t,x)dx1++bd(t,x)dxd (124)

the initial value X(t,x)=0 as prod(t,x)=0 and the initial density function is p(u;t,x)=δ(u) as prod(t,x)=0. Then, the density function is as follows:

p(u;t,x)=e12β(t,x)2πβ(t,x)ev22β(t,x)u=sinv (125)

Proof. 

Depending on the conditions, it is easy to obtain the Fokker–Planck equations as

tp(u;t,x)=b0(t,x)22u21u2p(u;t,x)3b0(t,x)2uup(u;t,x)xkp(u;t,x)=bk(t,x)22u21u2p(u;t,x)3bk(t,x)2uup(u;t,x)k=1,2,,d (126)

By implementing the transformation with u=sinv and defining p˜(v;t,x)=p(u(v;t,x)), the equations can be restructured as

tp˜(v;t,x)=b0(t,x)22v2p˜(v;t,x)+b0(t,x)2p˜(v;t,x)xkp˜(v;t,x)=bk(t,x)22v2p˜(v;t,x)+bk(t,x)2p˜(v;t,x)k=1,2,,d (127)

with the solution

p˜(v;t,x)=e120tb0(s,x)ds2π0tb0(s,x)dsev220tb0(s,x)dsp˜(v;t,x)=e120xkbk(t,x)dxk2π0xkbk(t,x)dxkev220xkbk(t,x)dxkk=1,2,,d (128)

Recall that a continuously differential function β(t,x) exists such that

dβ(t,x)=b0(t,x)dt+b1(t,x)dx1++bd(t,x)dxd (129)

we can derive the probability density function

p˜(v;t,x)=e12β(t,x)2πβ(t,x)ev22β(t,x) (130)

then,

p(u;t,x)=e12β(t,x)2πβ(t,x)ev22β(t,x)u=sinv (131)

Remark 2.

Similar to the discussion in Remark 1, we can obtain the Itô processes corresponding to the Fokker–Planck equations in Theorem 6

tX(t,x)=32b0(t,x)X(t,x)+b0(t,x)1X2(t,x)WtxkX(t,x)=32bk(t,x)X(t,x)+bk(t,x)1X2(t,x)Wk (132)

k=1,2,,d. In fact, this random field can be solved with a sinusoidal transformation, and the corresponding probability density function can be obtained. Although random field (132) has not yet found its application scenario, it gives us ideas for constructing different forms on space–time random fields in the future.

From density function (125), if we consider different β5(t,x) and β6(t,x), we can obtain density functions p(5)(u;t,x) and p(6)(u;t,x); then, we can obtain the Jeffreys divergence and generalized Fisher information

JDP(5)(t,x),P(5)(s,y)=1β5(t,x)β5(s,y)2β5(t,x)β5(s,y)β5(t,x)β5(s,y)2 (133)

and

FI1(t)P(5)(t,x)=12β5(t,x)2β52(t,x)b0(5)(t,x)2FI1(xk)P(5)(t,x)=12β5(t,x)2β52(t,x)bk(5)(t,x)2 (134)

and then the quotients

JDP(5)(t,x),P(5)(s,x)|ts|2=1β5(t,x)β5(s,x)2β5(t,x)β5(s,x)β5(t,x)β5(s,x)ts2JDP(5)(t,x),P(5)(t,x˜(k))|xkxk|2=1β5(t,x)β5(t,x˜(k))2β5(t,x)β5(t,x˜(k))β5(t,x)β5(t,x˜(k))xkxk2 (135)

Obviously, we can easily obtain

JDP(5)(t,x),P(5)(s,x)|ts|2FI1(t)P(5)(t,x)=1β5(t,x)β5(s,x)12β5(t,x)β5(t,x)β5(s,x)β5(t,x)β5(s,x)b0(5)(t,x)ts2JDP(5)(t,x),P(5)(t,x˜(k))|xkxk|2FI1(xk)P(5)(t,x)=1β5(t,x)β5(t,x˜(k))12β5(t,x)β5(t,x)β5(t,x˜(k))β5(t,x)β5(t,x˜(k))bk(5)(t,x)xxk2 (136)

for k=1,2,,d. Without a loss of generality, the result (136) corroborates Theorem 2.

Furthermore, if we consider different β5(t,x) and β6(t,x) in the density function (125), denoted as p(5)(u;t,x) and p(6)(u;t,x), we can obtain the generalized Fisher divergence at the same space–time points

FDbk(5),bk(6)P(5)(t,x)P(6)(t,x)=β5(t,x)β6(t,x)bk(6)β5(t,x)bk(5)β6(t,x)2β52(t,x)β62(t,x)β5(t,x)β52(t,x)+β6(t,x)β62(t,x) (137)

with the remainder terms

RkP(5)(t,x)P(6)(t,x)=bk(5)(t,x)bk(6)(t,x) (138)

for k=0,1,2,,d. Then, the generalized De Bruijn identities are as follows:

tJDP(5)(t,x),P(6)(t,x)=12FDb0(5),b0(6)P(5)(t,x)P(6)(t,x)b0(5)(t,x)b0(6)(t,x)xkJDP(5)(t,x),P(6)(t,x)=12FDbk(5),bk(6)P(5)(t,x)P(6)(t,x)bk(5)(t,x)bk(6)(t,x)k=1,2,,d (139)

5. Conclusions

In this paper, we provide a generalization of the classical definitions of entropy, divergence, and Fisher information and derive these measures on a space–time random field. In addition, the Fokker–Planck Equations (43) for the space–time random field and density functions are obtained. Moreover, we obtain the Jeffreys divergence of a space–time random field at different space–time positions, and we obtain the approximation of the ratio of Jeffreys divergence to the square of space–time coordinate difference to the generalized Fisher information (54). Additionally, we use the Jeffreys divergence on two space–time random fields from the same type but different parameters Fokker–Planck equations, to obtain generalized De Bruijn identities (61). Finally, we give three examples of Fokker–Planck equations, with their solutions, to calculate the corresponding Jeffreys divergence, generalized Fisher information, and Fisher divergence and obtain the De Bruijn identities. These results encourage further research into the entropy divergence of space–time random fields, which advances the pertinent fields of information entropy, Fisher information, and De Bruijn identities.

Acknowledgments

The author would like to thank Pingyi Fan, Zhanjie Song, Ying Li, and Yumeng Song for providing relevant references and helpful discussions on topics related to this work.

Abbreviations

The following abbreviations are used in this manuscript:

KL Kullback–Leibler divergence
FI Fisher information
CFI Cross–Fisher information
FD Fisher divergence
sFD Symmetric Fisher divergence
JD Jeffreys divergence

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declare no conflict of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Risken H. The Fokker–Planck Equation: Methods of Solution and Applications. Springer; Berlin/Heidelberg, Germany: 1984. [Google Scholar]
  • 2.Stam A.J. Some inequalities satisfied by the quantities of information of Fisher and Shannon. Inf. Control. 1959;2:101–112. doi: 10.1016/S0019-9958(59)90348-1. [DOI] [Google Scholar]
  • 3.Barron A.R. Entropy and the central limit theorem. Ann. Probab. 1986;14:336–342. doi: 10.1214/aop/1176992632. [DOI] [Google Scholar]
  • 4.Johnson O. Information Theory and the Central Limit Theorem. Imperial College Press; London, UK: 2004. [Google Scholar]
  • 5.Guo D. Relative entropy and score function: New information estimation relationships through arbitrary additive perturbation; Proceedings of the 2009 IEEE International Symposium on Information Theory; Seoul, Republic of Korea. 28 June–3 July 2009; pp. 814–818. [Google Scholar]
  • 6.Toranzo I.V., Zozor S., Brossier J.-M. Generalization of the De Bruijn Identity to General ϕ-Entropies and ϕ-Fisher Informations. IEEE Trans. Inform. Theory. 2018;64:6743–6758. doi: 10.1109/TIT.2017.2771209. [DOI] [Google Scholar]
  • 7.Kharazmi O., Balakrishnan N. Cumulative residual and relative cumulative residual Fisher information and their properties. IEEE Trans. Inform. Theory. 2021;67:6306–6312. doi: 10.1109/TIT.2021.3073789. [DOI] [Google Scholar]
  • 8.Kolmogorov A.N. The local structure of turbulence in incompressible viscous fluid for very large Reynolds numbers. Dokl. Akad. Nauk SSSR. 1941;30:299–303. [Google Scholar]
  • 9.Kolmogorov A.N. On the degeneration of isotropic turbulence in an incompressible viscous flu. Dokl. Akad. Nauk SSSR. 1941;31:538–542. [Google Scholar]
  • 10.Kolmogorov A.N. Dissipation of energy in isotropic turbulence. Dokl. Akad. Nauk SSSR. 1941;32:19–21. [Google Scholar]
  • 11.Yaglom A.M. Some classes of random fields in n-dimensional space, related to stationary random processes. Theory Probab. Its Appl. 1957;2:273–320. doi: 10.1137/1102021. [DOI] [Google Scholar]
  • 12.Yaglom A.M. Correlation Theory of Stationary and Related Random Functions. Volume I: Basic Results. Springer; New York, NY, USA: 1987. [Google Scholar]
  • 13.Yaglom A.M. Correlation Theory of Stationary and Related Random Functions. Volume II: Supplementary Notes and References. Springer; Berlin, Germany: 1987. [Google Scholar]
  • 14.Bowditch A., Sun R. The two-dimensional continuum random field Ising model. Ann. Probab. 2022;50:419–454. doi: 10.1214/21-AOP1536. [DOI] [Google Scholar]
  • 15.Bailleul I., Catellier R., Delarue F. Propagation of chaos for mean field rough differential equations. Ann. Probab. 2021;49:944–996. doi: 10.1214/20-AOP1465. [DOI] [Google Scholar]
  • 16.Wu L., Samorodnitsky G. Regularly varying random fields. Stoch. Process Their Appl. 2020;130:4470–4492. doi: 10.1016/j.spa.2020.01.005. [DOI] [Google Scholar]
  • 17.Koch E., Dombry C., Robert C.Y. A central limit theorem for functions of stationary max-stable random fields on Rd. Stoch. Process Their Appl. 2020;129:3406–3430. doi: 10.1016/j.spa.2018.09.014. [DOI] [Google Scholar]
  • 18.Ye Z. Ph.D. Dissertation. Cornell University; Ithaca, NY, USA: 1989. On Entropy and ε-Entropy of Random Fields. [Google Scholar]
  • 19.Ye Z., Berger T. A new method to estimate the critical distortion of random fields. IEEE Trans. Inform. Theory. 1992;38:152–157. doi: 10.1109/18.108261. [DOI] [Google Scholar]
  • 20.Ye Z., Berger T. Information Measures for Discrete Random Fields. Science Press; Beijing, China: New York, NY, USA: 1998. [Google Scholar]
  • 21.Ye Z., Yang W. Random Field: Network Information Theory and Game Theory. Science Press; Beijing, China: 2023. (In Chinese) [Google Scholar]
  • 22.Ma C. Stationary random fields in space and time with rational spectral densities. IEEE Trans. Inform. Theory. 2007;53:1019–1029. doi: 10.1109/TIT.2006.890721. [DOI] [Google Scholar]
  • 23.Hairer M. A theory of regularity structures. Invent. Math. 2014;198:269–504. doi: 10.1007/s00222-014-0505-4. [DOI] [Google Scholar]
  • 24.Hairer M. Solving the KPZ equation. Ann. Math. 2013;178:559–664. doi: 10.4007/annals.2013.178.2.4. [DOI] [Google Scholar]
  • 25.Kremp H., Perkowski N. Multidimensional SDE with distributional drift and Lévy noise. Bernoulli. 2022;28:1757–1783. doi: 10.3150/21-BEJ1394. [DOI] [Google Scholar]
  • 26.Beeson R., Namachchivaya N.S., Perkowski N. Approximation of the filter equation for multiple timescale, correlated, nonlinear systems. SIAM J. Math. Anal. 2022;54:3054–3090. doi: 10.1137/20M1379265. [DOI] [Google Scholar]
  • 27.Song Z., Zhang J. A note for estimation about average differential entropy of continuous bounded space–time random field. Chin. J. Electron. 2022;31:793–803. doi: 10.1049/cje.2021.00.213. [DOI] [Google Scholar]
  • 28.Kramers H.A. Brownian motion in a field of force and the diffusion model of chemical reactions. Physica. 1940;7:284–304. doi: 10.1016/S0031-8914(40)90098-2. [DOI] [Google Scholar]
  • 29.Moyal J.E. Stochastic processes and statistical physics. J. R. Stat. Soc. Ser. B Stat. Methodol. 1949;11:150–210. doi: 10.1111/j.2517-6161.1949.tb00030.x. [DOI] [Google Scholar]
  • 30.Shannon C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948;27:379–423, 623–656. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
  • 31.Neeser F.D., Massey J.L. Proper complex random processes with applications to information theory. IEEE Trans. Inform. Theory. 1991;39:1293–1302. doi: 10.1109/18.243446. [DOI] [Google Scholar]
  • 32.Ihara S. Information Theory-for Continuous Systems. World Scientific; Singapore: 1993. [Google Scholar]
  • 33.Gray R.M. Entropy and Information Theory. Springer; Boston, MA, USA: 2011. [Google Scholar]
  • 34.Bach F. Information Theory With Kernel Methods. IEEE Trans. Inform. Theory. 2023;69:752–775. doi: 10.1109/TIT.2022.3211077. [DOI] [Google Scholar]
  • 35.Kullback S., Leibler R.A. On information and sufficiency. Ann. Math. Stat. 1951;22:79–86. doi: 10.1214/aoms/1177729694. [DOI] [Google Scholar]
  • 36.Jeffreys H. An invariant form for the prior probability in estimation problems. Proc. R. Soc. Lond. A. 1946;186:453–461. doi: 10.1098/rspa.1946.0056. [DOI] [PubMed] [Google Scholar]
  • 37.Fuglede B., Topsøe F. Jensen-Shannon divergence and Hilbert space embedding; Proceedings of the IEEE International Symposium on Information Theory (ISIT); Chicago, IL, USA. 27 June–2 July 2004; [Google Scholar]
  • 38.Rényi A. Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics. Volume 1. University of California Press; Berkeley, CA, USA: 1961. On measures of entropy and information; pp. 547–561. [Google Scholar]
  • 39.She R., Fan P., Liu X.-Y., Wang X. Interpretable Generative Adversarial Networks With Exponential Function. IEEE Trans. Signal Process. 2021;69:3854–3867. doi: 10.1109/TSP.2021.3089285. [DOI] [Google Scholar]
  • 40.Liu S., She R., Zhu Z., Fan P. Storage Space Allocation Strategy for Digital Data with Message Importance. Entropy. 2020;22:591. doi: 10.3390/e22050591. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.She R., Liu S., Fan P. Attention to the Variation of Probabilistic Events: Information Processing with Message Importance Measure. Entropy. 2019;21:439. doi: 10.3390/e21050439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Wan S., Lu J., Fan P., Letaief K.B. Information Theory in Formation Control: An Error Analysis to Multi-Robot Formation. Entropy. 2018;20:618. doi: 10.3390/e20080618. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.She R., Liu S., Fan P. Recognizing Information Feature Variation: Message Importance Transfer Measure and Its Applications in Big Data. Entropy. 2018;20:401. doi: 10.3390/e20060401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Nielsen F. An Elementary Introduction to Information Geometry. Entropy. 2020;22:1100. doi: 10.3390/e22101100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Nielsen F. On the Jensen–Shannon Symmetrization of Distances Relying on Abstract Means. Entropy. 2019;21:485. doi: 10.3390/e21050485. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Nielsen F., Nock R. Generalizing skew Jensen divergences and Bregman divergences with comparative convexity. IEEE Signal Process. Lett. 2017;24:1123–1127. doi: 10.1109/LSP.2017.2712195. [DOI] [Google Scholar]
  • 47.Furuichi S., Minculete N. Refined Young Inequality and Its Application to Divergences. Entropy. 2021;23:514. doi: 10.3390/e23050514. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Pinele J., Strapasson J.E., Costa S.I. The Fisher-Rao Distance between Multivariate Normal Distributions: Special Cases, Bounds and Applications. Entropy. 2020;22:404. doi: 10.3390/e22040404. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Reverter F., Oller J.M. Computing the Rao distance for Gamma distributions. J. Comput. Appl. Math. 2003;157:155–167. doi: 10.1016/S0377-0427(03)00387-X. [DOI] [Google Scholar]
  • 50.Pawula R.F. Generalizations and extensions of the Fokker–Planck-Kolmogorov equations. IEEE Trans. Inform. Theory. 1967;13:33–41. doi: 10.1109/TIT.1967.1053955. [DOI] [Google Scholar]
  • 51.Pawula R.F. Approximation of the linear Boltzmann equation by the Fokker–Planck equation. Phys. Rev. 1967;162:186–188. doi: 10.1103/PhysRev.162.186. [DOI] [Google Scholar]
  • 52.Khoshnevisan D., Shi Z. Brownian Sheet and Capacity. Ann. Probab. 1999;27:1135–1159. doi: 10.1214/aop/1022677442. [DOI] [Google Scholar]
  • 53.Revuz D., Yor M. Continuous Martingales and Brownian Motion. 2nd ed. Springer; New York, NY, USA: 1999. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES