Skip to main content
Entropy logoLink to Entropy
. 2024 Apr 28;26(5):370. doi: 10.3390/e26050370

Intrinsic Information-Theoretic Models

D Bernal-Casas 1,*, J M Oller 1
Editors: Rosario Lo Franco1, Federico Holik1, Gustavo Martín Bosyk1, Ana Majtey1
PMCID: PMC11119449  PMID: 38785619

Abstract

With this follow-up paper, we continue developing a mathematical framework based on information geometry for representing physical objects. The long-term goal is to lay down informational foundations for physics, especially quantum physics. We assume that we can now model information sources as univariate normal probability distributions N (μ, σ0), as before, but with a constant σ0 not necessarily equal to 1. Then, we also relaxed the independence condition when modeling m sources of information. Now, we model m sources with a multivariate normal probability distribution Nm(μ,Σ0) with a constant variance–covariance matrix Σ0 not necessarily diagonal, i.e., with covariance values different to 0, which leads to the concept of modes rather than sources. Invoking Schrödinger’s equation, we can still break the information into m quantum harmonic oscillators, one for each mode, and with energy levels independent of the values of σ0, altogether leading to the concept of “intrinsic”. Similarly, as in our previous work with the estimator’s variance, we found that the expectation of the quadratic Mahalanobis distance to the sample mean equals the energy levels of the quantum harmonic oscillator, being the minimum quadratic Mahalanobis distance at the minimum energy level of the oscillator and reaching the “intrinsic” Cramér–Rao lower bound at the lowest energy level. Also, we demonstrate that the global probability density function of the collective mode of a set of m quantum harmonic oscillators at the lowest energy level still equals the posterior probability distribution calculated using Bayes’ theorem from the sources of information for all data values, taking as a prior the Riemannian volume of the informative metric. While these new assumptions certainly add complexity to the mathematical framework, the results proven are invariant under transformations, leading to the concept of “intrinsic” information-theoretic models, which are essential for developing physics.

Keywords: information geometry, fisher’s information, riemannian manifolds, schrödinger’s equation, principle of minimum fisher’s information, quantum harmonic oscillator, bayes’ theorem

1. Introduction

In this work, we continue developing the mathematical framework introduced in [1] by implementing some variations to better account for reality. In particular, we model information sources as univariate normal probability distributions N (μ, σ0) as before but with a constant σ0 not necessarily equal to 1. We also relax the independence condition when modeling m sources of information. Thus, we model m-dependent sources with a multivariate normal probability distribution Nm (μ, Σ0) with a constant variance–covariance matrix Σ0 not necessarily diagonal, i.e., with covariance values different than 0, which leads to the concept of modes rather than sources when finding the solutions.

As in our initial work, the mathematical approach departs from the supposition that physical objects are information-theoretic in origin, an idea that has recurrently appeared in physics. In the following mathematical developments, we discover that the approach is importantly “intrinsic”, giving rise to the paper’s title, which is the main feature we want to emphasize in this study. In other words, regardless of how we parametrize the modeling, the approach’s inherent properties, for example, energy levels, remain the same, irrespective of updating the framework with the above modifications.

This entire work builds upon this finding, which makes it ideal for studying the properties of information representation and developing physics, and our modifications can further improve the framework’s accuracy and applicability to real-world scenarios. The long-term goal is to provide models to explain the “pre-physics” stage from which everything may emerge. We refer to the initial preprocessing of the source data information which is performed, in principle, by our sensory systems or organs. Therefore, the research in this follow-up paper may significantly contribute to the field and potentially guide future work in the area.

2. Mathematical Framework

The plan of this section, which we divide into eleven subsections for didactic purposes, is the following. In Section 2.1, we outline modeling a single source with a single sample and the derivation of Fisher’s information and the Riemannian manifold. In Section 2.2, we describe modeling a single source with n samples. Section 2.3 is devoted to analyzing the stationary states of a single source with n samples in the Riemannian manifold. In Section 2.4, we present the solutions of the stationary states in our formalism. In Section 2.5, we compute the probability density function, the mean quadratic Mahalanobis distance, and the “intrinsic” Cramér–Rao lower bound for a single source with n samples. An extension of this approach to m independent sources is conducted in Section 2.6 to compute the global probability density function at the ground-state level. In Section 2.7, we outline modeling m-dependent sources of a single sample, Fisher’s information, and the Riemannian manifold. Section 2.8 describes m-dependent sources of n samples. In Section 2.9, we analyze the stationary states of m-dependent sources of n samples in the Riemannian manifold. Section 2.10 is devoted to finding the solutions. Finally, in Section 2.11, we use Bayes’ theorem to obtain the posterior probability density function.

2.1. A Single Source with a Single Sample: The Fisher’s Information, the Riemannian Manifold, and the Quadratic Mahalanobis Distance

We start our mathematical description by modeling a single source with a univariate normal probability distribution N(μ,σ0) where σ0>0 is a known constant. This is a well-known parametric statistical model in which unidimensional parameter space may be identified with the real line, i.e., Θ=R. We can compute all the relevant quantities relevant to our purpose. For a single sample, the univariate normal density (with respect to the Lebesgue measure), its natural logarithm, and the partial derivative with respect to the parameter μ are given by

f(x;μ)=12πσ0e12σ02(xμ)2, (1)
lnf=12ln(2πσ02)12σ02(xμ)2, (2)
lnfμ=xμσ02. (3)

From Equation (3), which is also called the score function, we can calculate Fisher’s information [2] for a single sample as

I(μ)=Eμlnfμ2=Eμxμσ022=1σ02Eμxμσ02=1σ02, (4)

since with the change z=xμσ0, we have that Eμxμσ02=z212πe12z2dz=1.

The Riemannian metric [3] from a single source with a single sample derived from the Fisher’s information (4) is a metric tensor whose covariant component, contravariant component, and determinant, respectively, are

g11(μ)=1σ02, (5)
g11(μ)=σ02, (6)
det(g(μ))=1σ02. (7)

The corresponding square of the Riemannian distance induced in the parametric space is the well-known quadratic Mahalanobis distance [4], i.e.,

dM2(μ2,μ1)=1σ02μ2μ12. (8)

The quadratic Mahalanobis distance will play a critical role in the next sections.

2.2. A Single Source with n Samples: The Fisher’s Information, the Riemannian Manifold, and the Square of the Riemannian Distance

If the source generates n samples, (x1,,xn), drawn independently from a univariate normal probability distribution (1), the likelihood of this n variate sample, its log-likelihood, and the scores, respectively, are

fn(x1,,xn;μ)=i=1nf(xi;μ)=12πσ02n2exp(12σ02i=1nxiμ2), (9)
lnfn=n2ln(2πσ02)12σ02i=1nxiμ2, (10)
lnfnμ=1σ02i=1nxiμ=nσ02x¯μ. (11)

Likewise, from Equation (11) we can calculate the Fisher’s information corresponding to an n size sample as

In(μ)=Eμlnfnμ2=nσ02Eμ(x¯μ)σ0/n2=nσ02. (12)

since x¯ follows a univariate normal distribution with mean equal to μ and variance equal to σ02n.

In other words

In(μ)=nI(μ), (13)

which shows the well-known additive property of the Fisher information for independent samples.

The Riemannian metric [3] from a single source with n samples derived from the Fisher’s information (12) is a metric tensor whose covariant component, contravariant component, and determinant, respectively, are

g˜11(μ)=nσ02, (14)
g˜11(μ)=σ02n, (15)
det(g˜(μ))=nσ02. (16)

The square of the Riemannian distance, ρ2, induced by the Fisher information matrix corresponding to a sample of arbitrary size n will be equal to n times the quadratic Mahalanobis distance (8), i.e.,

ρ2(μ2,μ1)=ndM2(μ2,μ1)=nσ02μ2μ12. (17)

2.3. Stationary States of a Single Source of n Samples in the Riemannian Manifold

To calculate the stationary states of a single source of n samples, we can invoke the principle of minimum Fisher information [5] or use the time-independent non-relativistic Schrodinger’s equation [6]. The two approaches have been demonstrated to be equivalent elsewhere [5]. The equation reads as follows

k2ψ(μ)+U(μ)ψ(μ)=λψ(μ), (18)

where U(μ) is a potential energy and k,λ>0. The solution must also satisfy limμψ(μ)=limμ+ψ(μ)=0 and ψ2(μ)dμ=1. For simplicity, we will write ψ instead of ψ(μ).

We can use the modulus square of the score function (11) as the potential energy, except for a constant term

(19a)lnfnμ2=g11(μ)lnfnμ2,(19b)=σ02nnσ02x¯μ2,(19c)=nσ02x¯μ2.

Alternatively, we can use as the potential energy the difference between the maximum of the log-likelihood attained by the sample, (x1,,xn), minus the log-likelihood at an arbitrary point μ, up to a proportionality constant. Since the likelihood is given by (9), we can rewrite it as

fn(x1,,xn;μ)=12πσ02n2exp(n2σ02sn2)exp(n2σ02(x¯μ)2), (20)

where x¯=1ni=1nxi and sn2=1ni=1n(xix¯)2. The supreme likelihood is obviously attained when μ=x¯, then, the previously mentioned potential will be

U(μ)lnfn(x1,,xn;x¯)lnfn(x1,,xn;μ)=n2σ02(x¯μ)2. (21)

This expression is up to a proportionality constant equal to (19). Thus, we may choose as the potential energy U(μ)=nCσ02x¯μ2 with C>0, and Equation (18) reads as:

k2ψ+nCσ02x¯μ2ψ=λψ, (22)

We compute the Laplacian in Equation (22) as:

(23a)2ψ=1|g(μ)|μ|g(μ)|g11(μ)ψμ,(23b)=σ0nμnσ0σ02nψμ,(23c)=σ02n2ψμ2=σ02nψ.

Inserting Equation (23) into Equation (22), we obtain:

kσ02nψ+nCσ02x¯μ2ψ=λψ, (24)

which is Schrödinger’s equation of the quantum harmonic oscillator [7].

2.4. Solutions of a Single Quantum Harmonic Oscillator in the Riemannian Manifold

Some steps now may seem obvious for those used to quantum mechanics. Considering that ψ has the following form:

ψ(μ)=γeη,withγ>0real,ηafunctionofμ. (25)

Equation (24) results:

(26a)kσ02nγeη(η)2+γeηη+nCσ02x¯μ2γeη=λγeη,(26b)kσ02n(η)2+η+nCσ02x¯μ2=λ.

Assuming a solution for η(μ) with the form

η(μ)=ξnx¯μ2,withξ>0, (27)

And inserting this expression into Equation (26) gives

kσ02n4ξ2n2x¯μ22ξn+nCσ02x¯μ2=λ, (28)

which implies that

4kσ02ξ2=Cσ02,2kσ02ξ=λ. (29)

In other words, k,C,λ,ξ can not be chosen arbitrarily because they have to satisfy these equations. For example, we can choose k=2n and C=12n, which forces ξ=14σ02, and λ=1n. Therefore, we can write

2σ02n2ψ+12σ02x¯μ2ψ=1nψ, (30)

whose solution is given by

ψ(μ)=γe14σ02nx¯μ2. (31)

With this configuration, we compute the normalization constant γ

1=ψ2(μ)dμ=γ2e12σ02nx¯μ2dμ=2γ2σ00e12nt2dt, (32)

where we used a first change of variable x¯μσ0=t. Now, using a second change of variable 12nt2=s, dt=2n12s12ds, Equation (32) writes as

1=2γ2σ00es2n12s12ds=γ22nσ00s12esds=γ22nπσ0. (33)

Isolating γ from Equation (33), we obtain γ=n2πσ0214. Therefore, Equation (31) reads as

ψ(μ)=n2πσ0214e14σ02nx¯μ2=ψ0(μ), (34)

which is the ground-state solution of the quantum harmonic oscillator problem, and the wave function for the ground-state. The solutions of the quantum harmonic oscillator involve Hermite Polynomials, which were introduced elsewhere [8,9]. In this way, we can prove, after some tedious but straightforward computations, that the wave function:

ψ1(μ)=γ1x¯μσ0e14σ02nx¯μ2,withγ1>0, (35)

is also a solution of

2σ02n2ψ+12σ02x¯μ2ψ=λ1ψ1, (36)

where λ1=3n is the energy of the first excited state, and γ1=n32πσ0214 is the normalization constant. With this representation, the λ’s (energy levels) are given by

λν=2n(ν+12)=Eν,withν=0,1,. (37)

Looking closely at Equation (37), we appreciate that the energy levels depend on two numbers, ν and n. The ground state at ν=0 has a finite energy E0=1n, and can become arbitrarily close to zero by massive sampling. Notably, the energy levels are independent of σ0. In other words, they do not depend on the informative parameters, leading to the concept of “intrinsic” information-theoretic models which will be discussed in greater detail later.

2.5. Probability Density Function of a Single Source of n Samples, Mean Quadratic Mahalanobis Distance, and Intrinsic Cramér–Rao Lower Bound

Assuming that the square modulus of the wave function can be interpreted as the probability density function:

ψ(μ)2=ψ*(μ)ψ(μ)=ρ(μ), (38)

we can compute the performance of the estimations of μ. For instance, we can calculate the expectation of the quadratic Mahalanobis distance (8) to the sample mean x¯ at the ground state (34), obtaining

Eμ,ρ0(μ)x¯μσ02=1nx¯μσ0/n212π(σ0/n)2e12(σ0/n)2x¯μ2dμ=1n. (39)

Likewise, we can compute the expectation of the quadratic Mahalanobis distance (8) to the sample mean x¯ at the first excited state, obtaining

Eμ,ρ1(μ)x¯μσ02=1n12π(σ0/n)2x¯μσ0/n4e12(σ0/n)2x¯μ2dμ=3n. (40)

The expectation of the quadratic Mahalanobis distances to the sample mean x¯ at the different states equal the quantum harmonic oscillator’s energy levels, i.e., this quantity is definitively quantized. Interestingly, the expectation of the quadratic Mahalanobis distance to the sample mean x¯ at the ground state (39) equals the intrinsic Cramér–Rao lower bound (ICRLB) for unbiased estimators

Eμx¯μσ02mn|m=1=1n, (41)

considering that we are modeling a single source of n samples with a single informative parameter μ, i.e., m=1. For further details, see [10].

2.6. m Independent Sources of n Samples and Global Probability Density Function

With m independent sources, each generating n samples, a finite set of m quantum harmonic oscillators may represent reality. Presuming independence of the sources of information, the “global” wave function (also called the collective wave function) can factor as the product of single wave functions. We can write the global wave function as

ψ(μ)=ψ(μ1,μ2,,μm)=ψ(μ1)ψ(μ2)ψ(μm), (42)

It constitutes a many-body system, and we may refer to the vector μ as the μ field.

For example, in the case of modeling two independent sources, the global wave function at the ground state will be the product of single wave functions, each of them at the ground state

(43a)ψ0(μ)=ψ0(μ1)ψ0(μ2),(43b)=n2πσ0214exp{14σ02nx¯1μ12}n2πσ0214exp{14σ02nx¯2μ22},(43c)=n2πσ0214n2πσ0214exp{14σ02nx¯1μ12+nx¯2μ22},(43d)=n2πσ0212exp{14σ02nx¯1μ1,x¯2μ2x¯1μ1x¯2μ2}.

We can generalize Equation (43) for having m independent sources. The global wave function is written as

ψ0(μ)=n2πσ02m4exp{14σ02nx¯μTx¯μ}. (44)

Using Equation (38), the probability density function is:

ρ0(μ)=n2πσ02m2exp{12σ02nx¯μTx¯μ}. (45)

2.7. m Dependent Sources of a Single Sample, the Fisher’s Information, the Riemannian Manifold, and the Quadratic Mahalanobis Distance

Consider now m possibly dependent sources, generating a multivariate sample of size 1, x. Now, although we have a finite set of m univariate quantum harmonic oscillators that can also represent reality, since these sources may be dependent, it is convenient to model this situation as a single m variate source with an m-variate random vector x following a m variate normal probability distribution Nm(μ,Σ0) where μRm and Σ0 is a known constant strictly positive definite m×m matrix, the covariance matrix of random vector x, i.e., cov(x)=Σ0>0. This is a well-known parametric statistical model in which their m-dimensional parameter space may be identified with Θ=Rm; for further details, see [11]. Identifying, as is customary, the points of the manifold Θ with their coordinates μ=(μ1,,μm), we can compute all the quantities relevant to our purpose. For a single sample, the m-variate normal density (with respect to the Lebesgue measure), its natural logarithm, and the partial derivative with respect to μi are given by

fm(x;μ)=(2π)m2det(Σ0)12exp(12(xμ)TΣ01(xμ)), (46)
lnfm=m2ln(2π)12ln(det(Σ0))12(xμ)TΣ01(xμ), (47)
lnfmμi=α=1mσiαxαμα. (48)

where, following a standard matrix calculus notation, σiα is the element located in row i and column α of the inverse covariance matrix Σ0.

The Fisher’s information matrix G=gij(μ) is a m×m matrix whose elements are

(49a)gij(μ)=Eμlnfmμilnfmμj,(49b)=Eμα=1mσiα(xαμα)β=1mσjβ(xβμβ,(49c)=α=1mβ=1mσiασjβEμ(xαμα)(xβμβ),(49d)=β=1mα=1mσiασαβσβj=β=1mδβiσβj=σij,

where we have taken into account the symmetry of σjβ, and that cov((xαμα),(xβμβ))=Eμ(xαμα)(xβμβ)=σαβ, or, in matrix form, G=Σ01.

It is well known that the Fisher information matrix is a second-order covariant tensor of the parameter space. It is positive definite and may be considered the metric tensor of this manifold, giving it the structure of the Riemannian manifold. To avoid confusion, we must emphasize that the subscripts or superscripts used to reference the variance and covariance matrix or its inverse do not have a tensor character, i.e., the components of the metric tensor gij(μ) are those of a covariant tensor, in tensor notation, they are written as subscripts and are equal to the components of the inverse variance–covariance matrix given in (49).

The Riemannian geometry induced by the Fisher information matrix in the parameter space is, in this case, Euclidean, and the square of the Riemannian distance, also known as the Rao distance, is the Mahalanobis distance given by

dM2(μ2,μ1)=(μ2μ1)TΣ01(μ2μ1). (50)

In this expression, the parameter space points are identified with their coordinates and written in matrix notation as m×1 column vectors.

All these results correspond to a multivariate with a sample size n=1.

2.8. m Dependent Sources of n Samples, the Fisher’s Information, the Riemannian Manifold, and the Square of the Riemannian Distance

If each of the m dependent sources generates n samples, (x1,,xn), drawn independently from a multivariate normal probability distribution (46), the likelihood distribution is

(51a)fm,n(x1,,xn;μ)=i=1n(2π)m2det(Σ0)12exp(12(xiμ)TΣ01(xiμ)),(51b)=(2π)mn2det(Σ0)n2exp(12i=1n(xiμ)TΣ01(xiμ)).

The summation term within the exponential function can be decomposed into two terms

(52a)i=1n(xiμ)TΣ01(xiμ)=i=1nTr(Σ01(xiμ)(xiμ)T),(52b)=TrΣ01i=1n(xiμ)(xiμ)T,(52c)=TrΣ01i=1n(xix¯+x¯μ)(xix¯+x¯μ)T,(52d)=TrΣ01i=1n(xix¯)(xix¯)T+i=1n(x¯μ)(x¯μ)T,(52e)=nTrΣ01Sn+ndM2(x¯,μ),

where x¯=(i=1nxi)/n, and Sn=(i=1n(xix¯)(xix¯)T)/n, i.e., the sample mean and the sample covariance matrix corresponding to this random sample is the quadratic Mahalanobis distance to the sample mean, x¯.

Inserting Equation (52) into Equation (51) results

fm,n(x1,,xn;μ)=(2π)mn2det(Σ0)n2expn2Tr{Σ01Sn}expn2dM2(x¯,μ). (53)

The log-likelihood distribution is

lnfm,n=mn2ln(2π)n2lndet(Σ0)n2Tr{Σ01Sn}n2dM2(x¯,μ), (54)

and the partial derivative of lnfm,n with respect to μα using standard classical notation for covariant derivatives and additionally using repeated index summation convention is

lnfm,nμα=lnfm,n,α=ngαβ(x¯βμβ). (55)

The corresponding Fisher information matrix and the square of Riemannian distance for a sample of size n will be the above Equations (49) and (50) multiplied by n.

2.9. Stationary States of m-Dependent Sources of n Samples in the Riemannian Manifold

To calculate the stationary states, we can invoke the time-independent non-relativistic Schrodinger’s equation [6] as above. In the multivariate case, the wave equation reads as follows

k2ψ(μ)+U(μ)ψ(μ)=λψ(μ), (56)

where U(μ) is the potential energy and k,λ>0. The solution must also satisfy that vanish to infinity and Rmψ2(μ)dμ=1. For simplicity, we will write ψ instead of ψ(μ).

We can use the square of the norm of the log-likelihood gradient as the potential energy, except for a constant term. The components of the gradient of (lnfm,n), a contravariant vector field, observing that the inverse of the metric tensor corresponding to a sample of size n is given by 1ngγα since gγαgαβ=δβγ where δβγ is the Kronecker delta’s, will be given in classical notation by

(lnfm,n)γ=(lnfm,n),γ=1ngγα(lnfm,n),α=1ngγαngαβ(x¯βμβ)=(x¯γμγ), (57)

and, therefore, the square of the norm of the log-likelihood gradient will be

lnfm,n2=ngγβ(lnfm,n),γ(lnfm,n),β=ngγβ(x¯γμγ)(x¯βμβ)=ndM2(x¯,μ). (58)

Alternatively, we can use the difference between the log-likelihood at an arbitrary point μ0=x¯ minus the log-likelihood attained by the sample as the potential

U(μ)lnfm,n(x¯)lnfm,n(μ)=n2dM2(x¯,μ), (59)

which is up to a proportionality constant equal to (58). Thus, Equations (58) and (59) suggest to take as the potential energy U(μ)=nCdM2(x¯,μ) with C>0. In this way, Equation (56) reads as

k2ψ+nCdM2(x¯,μ)ψ=λψ. (60)

To proceed, we must compute the Laplacian in Equation (60). If g=det(G), with G defined in (49), i.e., the determinant corresponding to the tensor of the information metric for samples of size n=1, for a sample of arbitrary size n that determinant will be equal to nmg. In this way, the Laplacian of a function ψ will be given by

2ψ=1nmgμinmg1ngijψμj=1ngij2ψμiμj, (61)

where we have used repeated index summation convention. For further details about this formula, see, for instance, [12]. Notice that gij equals the variance–covariance matrix Σ0. Moreover, i m×m matrix Ψμ=2ψμiμjm×m, which is the Hessian matrix of ψ under the coordinates μ=(μ1,,μm), Equation (61) can be written as

2ψ=1nTrΣ0Ψμ. (62)

Inserting Equation (62) into (60), we obtain

knTrΣ0Ψμ+nCdM2(μ,x¯n)ψ=λψ, (63)

which is the Schrödinger’s equation of m-coupled quantum harmonic oscillators.

2.10. Solutions of m-Coupled Quantum Harmonic Oscillators in the Riemannian Manifold

We observe that both (58) and (61) are invariant under coordinate changes on Θ=Rn. Therefore, Equation (63) will remain invariant under these changes, particularly linear ones.

Since Σ0 is a symmetric m×m matrix which diagonalizes on an orthonormal basis, it can be written as Σ0=UDUT, where U is an orthogonal m×m matrix and D is a diagonal m×m matrix D=diag(γ12,,γm2), i=1,,m, where γα>0 are the eigenvalues of the square root of the variance–covariance matrix Σ0.

Thus, by introducing the change of coordinates η=UTμ and y¯=UTx¯, the metric tensor becomes diagonal, i.e., G^=D1. Taking this coordinate change into account, the Equation (58) becomes

ndM2(x¯,μ)=ndM2(y¯,η)=i=1m1γi2(y¯iηi)2. (64)

Moreover, if we define the symmetric m×m matrix Ψη=2Ψηiηjm×m, which is the Hessian matrix of ψ under the new coordinates η=(η1,,ηm), the Equation (61) becomes

2ψ=1nTrΣ0Ψμ=1nTrDΨη. (65)

Making use of Equation (64) and Equation (65) in Equation (63), we obtain

knTr(DΨη)+nCdM2(y¯,η)ψ=λψ, (66)

which is the Schrödinger’s equation of the m-decoupled quantum harmonic oscillators. If we choose k=2n and C=12n, Equation (66) can be written as

i=1m2γi2n22ψ(ηi)2+12γi2(y¯iηi)2ψ=λψ. (67)

Additionally, if we define λ=i=1mλα with λα>0, we may write

i=1m2γi2n22ψ(ηi)2+12γi2(y¯iηi)2ψλαψ=0. (68)

Note, that if ψα is a non–trivial solution of

2γα2n22ψα(ηα)2+12γα2(y¯αηα)2ψα=λαψα,α=1,,m. (69)

then, Equation (66) admits a solution of the form

ψ(η1,,ηm)=α=1mψα(ηα). (70)

Each of the Equations in (69) admits infinite solutions for different values of λα, as in our previous work [1]. More specifically, it admits solutions for

λα,ν=2n(ν+12),νN. (71)

In particular, for ν=0, we have λα,0=1n and the wave function for the ground-state is

ψα,0(ηα)=n2πγα214e14γα2ny¯αηα2, (72)

Then, the global wave function at the ground state will be

ψ0(η1,,ηm)=α=1mψα,0(ηα), (73)

with λ=α=1m1n=mn, which is the intrinsic Cramér–Rao lower bound (ICRLB) for m sources of information of n samples, with each source being modeled with an informative parameter μ, i.e., a total of m informative parameters. For further details, see [10]. The global probability density function at the ground state can be written as

(74a)ρ0(η1,,ηm)=ψ0(η1,,ηm)2=α=1mψα,02(ηα),(74b)=n2πm2α=1mγα212expn2α=1m1γα2y¯αηα2,(74c)=n2πm2det(D)12exp(n2y¯αηαTD1y¯αηα).

Since η=UTμ and y¯=UTx¯, where U is an orthonormal m×m matrix and, therefore, |detU|=1, we can express Equation (74) as a function of the original coordinates

ρ0(μ1,,μm)=n2πm2det(Σ0)12exp(n2(x¯μ)TΣ01(x¯μ)) (75)

However, there are many other solutions in (69) considering different values of ν in (71). It is well-known that the solutions of the quantum harmonic oscillator involve Hermite Polynomials, which were introduced elsewhere [8,9]. In particular, for ν=1, we will have λα,1=3n and the wave function at the first-excited state will be

ψα,1(ηα)=n32πγα214y¯αηαγαe14γα2ny¯αηα2. (76)

We can obtain other solutions via the Hermite polynomials, representing excited states of the quantum harmonic oscillator. For instance, we may obtain the solution ψα,ν for each of the sources α=1,,m and for each energy level ν=0,,k. Combining the m sources with the k+1 energy levels, we can build up all possible solutions and, therefore, obtain up to (k+1)m different solutions

ψ(η1,,ηm)=α=1mψα,ϵα(ηα), (77)

where ϵα{0,1,,k} and λ=α=1m2nϵα+12=mn+2nα=1mϵα, the total energy of m oscillators, such that mnλ2mkn.

2.11. Bayesian Framework and Posterior Probability Density Function

Regardless of having independent or dependent sources of information, we can compute the posterior probability distribution calculated from the sources of information for all data values using Bayes’theorem [13], taking the Riemannian volume of the metric as a prior. This measure is Jeffrey’s prior distribution on the parameter, and it can be considered in some way an objective, or at least a reference choice for a prior distribution [14].

Considering Equation (49), the Riemannian volume det(G) is constant in Θ. Then, taking into account the likelihood probability distribution (51), the posterior probability distribution fm,n(μ;x1,,xn) based on Jeffrey’s prior is equal to

(78a)fm,n(μ;x1,,xn)fm,n(x1,,xn;μ)g,(78b)exp(12ndM2(x¯,μ)),(78c)=n2πm2det(Σ0)12exp(n2dM2(x¯,μ)),(78d)=n2πm2det(Σ0)12exp(n2(x¯μ)TΣ01(x¯μ))

This posterior probability density function coincides with the global probability density function at the ground state (75). Precisely, the probability density function of m quantum harmonic oscillators at the ground state given by the square of the wave function coincides with the Bayesian posterior obtained from m sources of information for all data values using the improper Jeffrey’s prior. This unexpected and exciting result reveals a plausible relationship between energy and Bayes’ theorem.

3. Discussion

This paper aimed to expand and refine the mathematical framework initially presented in [1]. We made specific adjustments to the approach, enabling us to consider real-world scenarios more thoroughly. As we continued with our work, we came to appreciate the “intrinsic” nature of the modeling, which we believe is a crucial aspect of our study. Our ultimate objective was to improve upon the foundation established in the previous paper and create an even more robust and accurate framework.

First, we extended the approach by modeling a single source of information with a univariate normal probability distribution N(μ,σ0), as before, but with a constant σ0 not necessarily equal to 1. We calculated the stationary states in the Riemannian manifold by invoking Schrödinger’s equation to discover that the information could be broken into quantum harmonic oscillators as before but with the energy levels being independent of σ0, an unexpected but relevant result that motivated us to continue exploring this field.

This primitive result led us to title the work “Intrinsic information-theoretic models”, which asserts that the critical features of our modeling process, such as the energy levels, remain independent of the parametrization used and invariant under coordinate changes. This notion of invariance is significant because it implies that the same model can be applied across different parameterizations, allowing for greater consistency and generalizability. Furthermore, this approach can lead to a more robust and reliable modeling process, as it reduces the impact of specific parameter choices on the final model output. As such, the notion of “intrinsic” information-theoretic models has the potential to improve modeling accuracy and reliability significantly.

Similar to our previous study [1], we evaluated the performance of the estimation of the parameter μ. Instead of calculating the estimator’s variance, we used the expectation of the quadratic Mahalanobis distance to the sample mean to discover that equals the energy levels of the quantum harmonic oscillator, being the minimum quadratic Mahalanobis distance at the minimum energy level of the oscillator. Interestingly, we demonstrated that quantum harmonic oscillators reach the “intrinsic” Cramér–Rao lower bound on the quadratic Mahalanobis distance at the lowest energy level.

Then, we modeled m independent sources of information and computed the global density function at the ground state as an example. Essentially, we modeled sources with a multivariate normal probability distribution Nm(μ,Σ0), with a variance–covariance matrix Σ0 different than the identity matrix of m-dimension, Im, but being diagonal initially to describe the independence of the sources of information.

We advanced the mathematical approach by modeling m dependent sources of information with a variance–covariance matrix Σ0 not necessarily diagonal, depicting dependent sources. This resulted in Schrödinger’s equation of m-coupled quantum harmonic oscillators. We could effectively decouple the oscillators through a coordinated transformation, thereby partitioning the information into independent modes. This enabled us to obtain the same energy levels, albeit now with respect to the modes, which further proves the “intrinsic” property of the mathematical framework.

Finally, as in our previous study, we showed that the global probability density function of a set of m quantum harmonic oscillators at the lowest energy level, calculated as the square modulus of the global wave function at the ground state, equals the posterior probability distribution calculated using Bayes’ theorem from the m sources of information for all data values, taking as a prior the Riemannian volume of the informative metric. This is true regardless of whether the sources are independent or dependent.

Apart from the mathematical discoveries detailed in this paper, this framework offers multiple alternatives that we are currently exploring. For example, the informational representation of statistical manifolds with Σ0 is unknown. Also, this approach can be generalized by exploring other statistical manifolds and depicting how physical observables such as space and time may emerge from linear and nonlinear transformations of a set of parameters of a specific statistical manifold. This way, the laws of physics, including time’s arrow, will appear afterward.

Moreover, several fascinating inquiries warrant further investigation. These involve delving into the relationship between energy and information already highlighted in our initial work. In addition, the very plausible connection between energy and Bayes’ theorem also deserves further exploration. By delving deeper into these topics, we may unlock even more insights into the universe’s fundamental nature and mathematical laws.

The updated framework presented in this study offers a more realistic approach by allowing the modeling of m-dependent sources. In real-world scenarios, information is often distributed over multiple sources that may not be entirely independent. By formulating the problem in terms of modes, we can obtain a solution or set of solutions for the proposed framework. This approach provides a valuable tool for solving complex problems that require a deeper understanding of the underlying dynamics.

Author Contributions

Conceptualization, D.B.-C. and J.M.O.; writing—original draft preparation, D.B.-C. and J.M.O.; writing—review and editing, D.B.-C. and J.M.O. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

The study did not require ethical approval.

Data Availability Statement

No new data were created.

Conflicts of Interest

The authors declare no conflicts of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Bernal-Casas D., Oller J.M. Information-Theoretic Models for Physical Observables. Entropy. 2023;25:1448. doi: 10.3390/e25101448. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Fisher R. On the mathematical foundations of theoretical statistics. Philos. Trans. R. Soc. Lond. Ser. Contain. Pap. Math. Phys. Character. 1922;222:309–368. [Google Scholar]
  • 3.Riemann B. Über die Hypothesen, Welche der Geometrie zu Grunde Liegen. (Mitgetheilt durch R. Dedekind) 1868. [(accessed on 15 July 2023)]. Available online: https://eudml.org/doc/135760.
  • 4.Mahalanobis P. On the generalized distance in Statistics. Proc. Nat. Inst. Sci. India. 1936;2:49–55. [Google Scholar]
  • 5.Frieden B. Science from Fisher Information: A Unification. 2nd ed. Cambridge University Press; Cambridge, UK: 2004. [Google Scholar]
  • 6.Schrödinger E. An Undulatory Theory of the Mechanics of Atoms and Molecules. Phys. Rev. 1926;28:1049–1070. doi: 10.1103/PhysRev.28.1049. [DOI] [Google Scholar]
  • 7.Schrödinger E. Quantisierung als Eigenwertproblem. II. Ann. Phys. 1926;79:489–527. doi: 10.1002/andp.19263840602. [DOI] [Google Scholar]
  • 8.Laplace P. Mémoires de la Classe des Sciences Mathématiques et Physiques de L’institut Impérial de France. Institut de France; Paris, France: 1811. Mémoire sur les Intégrales Définies et leur Application aux Probabilités, et Spécialement a la Recherche du Milieu Qu’il Faut Choisir Entre les Resultats des Observations; pp. 297–347. [Google Scholar]
  • 9.Hermite C. Sur un Nouveau Développement en Série de Fonctions. Académie des Sciences and Centre National de la Recherche Scientifique de France; Paris, France: 1864. [Google Scholar]
  • 10.Oller J.M., Corcuera J.M. Intrinsic Analysis of Statistical Estimation. Ann. Stat. 1995;23:1562–1581. doi: 10.1214/aos/1176324312. [DOI] [Google Scholar]
  • 11.Muirhead R. Aspects of Multivariate Statistical Theory. Wiley; Hoboken, NJ, USA: 1982. [DOI] [Google Scholar]
  • 12.Chavel I. Eigenvalues in Riemannian Geometry. Elsevier; Philadelphia, PA, USA: 1984. [DOI] [Google Scholar]
  • 13.Bayes T. An essay towards solving a problem in the doctrine of chances. Phil. Trans. R. Soc. Lond. 1763;53:370–418. doi: 10.1093/biomet/45.3-4.296. [DOI] [PubMed] [Google Scholar]
  • 14.Jeffreys H. An invariant form for the prior probability in estimation problems. Proc. R. Soc. Lond. Ser. Math. Phys. Sci. 1946;186:453–461. doi: 10.1098/rspa.1946.0056. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No new data were created.


Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES