Skip to main content
Springer logoLink to Springer
. 2023 Apr 5;86(5):69. doi: 10.1007/s00285-023-01895-8

Embeddability of centrosymmetric matrices capturing the double-helix structure in natural and synthetic DNA

Muhammad Ardiyansyah 1,, Dimitra Kosta 2, Jordi Roca-Lacostena 3
PMCID: PMC10076417  PMID: 37017794

Abstract

In this paper, we discuss the embedding problem for centrosymmetric matrices, which are higher order generalizations of the matrices occurring in strand symmetric models. These models capture the substitution symmetries arising from the double helix structure of the DNA. Deciding whether a transition matrix is embeddable or not enables us to know if the observed substitution probabilities are consistent with a homogeneous continuous time substitution model, such as the Kimura models, the Jukes-Cantor model or the general time-reversible model. On the other hand, the generalization to higher order matrices is motivated by the setting of synthetic biology, which works with different sizes of genetic alphabets.

Keywords: Evolutionary model, Embedding problem, Markov matrix, Centrosymmetric matrix

Introduction

Phylogenetics is the study of evolutionary relationships among biological entities, also known as taxa, that aims to infer the evolutionary history among them. In order to model evolution, we consider a directed acyclic graph, called a phylogenetic tree, depicting the evolutionary relationships amongst a selected set of taxa. Phylogenetic trees consist of vertices and edges. Vertices represent taxa, while edges between vertices represent the evolutionary processes between the taxa.

In order to describe the real evolutionary process along an edge of a phylogenetic tree, one often assumes that the evolutionary data occurred following a Markov process. A Markov process is a random process in which the future is independent of the past, given the present. Under this Markov process, transitions between n states given by conditional probabilities are presented in a n×n Markov matrix M, that is a square matrix whose entries are nonnegative and rows sum to one. A well-known problem in probability theory is the so-called embedding problem which was initially posed by Elfving (Elfving 1937). The embedding problem asks whether given a Markov matrix M, one can find a real square matrix Q with rows summing to zero and non-negative off-diagonal entries, such that M=exp(Q). The matrix Q is called a Markov generator.

In the complex setting, the embedding problem is completely solved by Higham (2008); a complex matrix A is embeddable if and only if A is invertible. However, as our motivation arises from molecular models of evolution we are interested in the embedding problem over the real numbers, so from now on we will denote by M a real Markov matrix. It was shown by Kingman (1962) that if an n×n real Markov matrix M is embeddable, then the matrix M has detM>0. Moreover, in the same work by Kingman it was shown that detM>0 is a necessary and sufficient condition for a 2×2 Markov matrix M to be embeddable. For 3×3 Markov matrices a complete solution of the embedding problem is provided in a series of papers (James 1973; Johansen 1974; Carette 1995; Chen and Chen 2011), where the characterisation of embeddable matrices depends on the Jordan decomposition of the Markov matrix. For 4×4 Markov matrices the embedding problem is completely settled in a series of papers (Casanellas et al. 2020a, 2023; Roca-Lacostena and Fernández-Sánchez 2018a), where similarly to the 3×3 case the full characterisation of embeddable matrices is distinguished into cases depending on the Jordan form of the Markov matrices.

For the general case of n×n Markov matrices, there are several results; some presenting necessary conditions (Elfving 1937; Kingman 1962; Runnenberg 1962), while others sufficient conditions (James 1973; Fuglede 1988; Goodman 1970; Davies et al. 2010) for embeddability of Markov matrices. Moreover, the embedding problem has been solved for special n×n matrices with a biological interest such as equal-input and circulant matrices (Baake and Sumner 2020), group-based models (Ardiyansyah et al. 2021) and time-reversible models (Jia 2016). Despite the fact that there is no theoretical explicit solution for the embeddability of general n×n Markov matrices, there are results (Casanellas et al. 2023) that enable us to decide whether a n×n Markov matrix with distinct eigenvalues is embeddable or not. This is achieved by providing an algorithm that outputs all Markov generators of such a Markov matrix (Casanellas et al. 2023; Roca-Lacostena 2021).

In this paper, we focus on the embedding problem for n×n matrices that are symmetric about their center and are called centrosymmetric matrices (see Definition 2). We also study a variation of the famous embedding problem called model embeddability, where apart from the requirement that the Markov matrix is the matrix exponential of a rate matrix, we additionally ask that the rate matrix follows the model structure. For instance, for centrosymmetric matrices, model embeddability means that the rate matrix is also centrosymmetric.

The motivation for studying centrosymmetric matrices comes from evolutionary biology, as the most general nucleotide substitution model when considering both DNA strands admits any n×n centrosymmetric Markov matrix as a transition matrix, where n is the even number of nucleotides. For instance, by considering the four natural nucleotides A-T, C-G we arrive at the strand symmetric Markov model, a well-known phylogenetic model whose substitution probabilities reflect the symmetry arising from the complementarity between the two strands that the DNA is composed of (see (Casanellas and Sullivant 2005)). In particular, a strand symmetric model for DNA must have the following equalities of probabilities in the root distribution:

πA=πTandπC=πG 1.1

and the following equalities of probabilities in the transition matrices (θij)

θAA=θTT,θAC=θTG,θAG=θTC,θAT=θTA,θCA=θGT,θCC=θGG,θCG=θGC,θCT=θGA.

Therefore, the corresponding transition matrices of this model are 4×4 centrosymmetric matrices, usually called strand symmetric Markov matrices in this context. In this article, we will use the terminology 4×4 centrosymmetric Markov matrix and strand symmetric Markov matrix interchangeably. In the strand symmetric model there are less restrictions on the way genes mutate from ancestor to child compared to other widely known molecular models of evolution. In fact, special cases of the strand symmetric model are the group-based phylogenetic models such as the Jukes-Cantor (JC) model, the Kimura 2-parameter (K2P) and Kimura 3-parameter (K3P) models. The algebraic structure of strand symmetric models was initially studied in (Casanellas and Sullivant 2005), where it was argued that strand symmetric models capture more biologically meaningful features of real DNA sequences than the commonly used group-based models, as for instance, in any group-based model, the stationary distribution of bases for a single species is always the uniform distribution, while computational evidence in (Yap and Pachter 2004) suggests that the stationary distribution of bases for a single species is rarely uniform, but must always satisfy the symmetries (1.1) arising from nucleotide complementarity, as assumed by the strand symmetric model.

In this article, we also explore higher order centrosymmetric matrices for which n>4, which is justified by the use of synthetic nucleotides. One of main goals of synthetic biology is to expand the genetic alphabet to include an unnatural or synthetic base pair. The more letters in a genetic system could possibly lead to an increased potential for retrievable information storage and bar-coding and combinatorial tagging (Benner and Sismour 2005). Naturally the four-letter genetic alphabet consists of just two pairs, A-T and G-C. In 2012, a genetic system comprising of three base pairs was introduced in (Malyshev et al. 2012). In addition to the natural base pairs, the third, unnatural or synthetic base pair 5SICS-MMO2 was proven to be functionally equivalent to a natural base pair. Moreover, when it is combined with the natural base pairs, 5SICS-MMO2 provides a fully functional six-letter genetic alphabet. Namely, six-letter genetic alphabets can be copied (Yang et al. 2007), polymerase chain reaction (PCR)-amplified and sequenced (Sismour et al. 2004; Yang et al. 2011), transcribed to six-letter RNA and back to six-letter DNA (Leal et al. 2015), and used to encode proteins with added amino acids (Bain et al. 1992). This biological importance and relevance of the above six-letter genetic alphabets motivates us to particularly study the 6×6 Markov matrices describing the probabilities of changing base pairs in the six-letter genetic system in Sect. 6. When considering both DNA strands, each substitution is observed twice due to the complementarity between both strands, and hence the resulting transition matrix is centrosymmetric.

Moreover there are other synthetic analogs to natural DNA which justify studying centrosymmetric matrices for n>6. For instance, hachimoji DNA is a synthetic DNA that uses four synthetic nucleotides B,Z,P,S in addition to the four natural ones A,C,G,T. With the additional four synthetic ones, hachimoji DNA forms four types of base pairs, two of which are unnatural: P binds with Z and B binds with S. The complementarity between both strands of the DNA implies that the transition matrix is centrosymmetric. Moreover, the research group responsible for the hachimoji DNA system had also studied a synthetic DNA analog system that used twelve different nucleotides, including the four found in DNA (see Yang et al. 2006). Although the biological models which motivate the study of centrosymmetric matrices in this paper require n to be an even number due to the double-helix structure of DNA, in Sect. 5, we include the case of n being odd for completeness.

Apart from embeddability, that is existence of Markov generators, it is also natural to ask about uniqueness of a Markov generator which is called the rate identifiability problem. Identifiability is a property which a model must satisfy in order for precise statistical inference to be possible. A class of phylogenetic models is identifiable if any two models in the class produce different data distributions. In this article, we further develop the results on rate identifiability of the Kimura two parameter model (Casanellas et al. 2020a) to study rate identifiability for strand symmetric models. We also show that there are embeddable strand symmetric Markov matrices with non identifiable rates, namely the Markov generator is not unique. Moreover, we show that strand symmetric Markov matrices are not generically identifiable, that is, there exists a positive measure subset of strand symmetric Markov matrices containing embeddable matrices whose rates are not identifiable.

This paper is organised as follows. In Sect. 2, we introduce the basic definitions and results on embeddability. In Sect. 3, we give a characterisation for a 4×4 centrosymmetric Markov matrix M with four distinct real nonnegative eigenvalues to be embeddable providing necessary and sufficient conditions in Theorem 2, while we also discuss their rate identifiability property in Proposition 3. Moreover in Sect. 4, using the conditions of our main result Theorem 2, we compute the relative volume of all 4×4 centrosymmetric Markov matrices relative to the 4×4 centrosymmetric Markov matrices with positive eigenvalues and Δ>0, as well as the relative volume of all 4×4 centrosymmetric Markov matrices relative to the 4×4 centrosymmetric Markov matrices with four distinct eigenvalues and Δ>0. We also compare the results on relative volumes obtained using our method with the algorithm suggested in Casanellas et al. (2023) to showcase the advantages of our method. In Sect. 5, we study higher order centrosymmetric matrices and motivate their use in Sect. 6 by exploring the case of synthetic nucleotides where the phylogenetic models admit 6×6 centrosymmetric mutation matrices. Finally, Sect.  7 discusses implications and possibilities for future work.

Preliminaries

In this section we will introduce the definitions and results that will be required throughout the paper. We will denote by Mn(K) the set of n×n square matrices with entries in the field K=RorC. The subset of non-singular matrices in Mn(K) will be denoted by GLn(K).

Definition 1

A Markov (or transition) matrix is a non-negative real square matrix with rows summing to one. A rate matrix is a real square matrix with rows summing to zero and non-negative off-diagonal entries.

In this paper, we are focusing on a subset of Markov matrices called centrosymmetric Markov matrices.

Definition 2

A real n×n matrix A=(ai,j) is said to be centrosymmetric (CS) if

ai,j=an+1-i,n+1-j

for every 1i,jn.

Definition 2 reveals that a CS matrix is nothing more than a square matrix which is symmetric about its center. This class of matrices has been previously studied, for instance, in (Aitken 2017, page 124) and Weaver (1985). Examples of CS matrices for n=5 and n=6, are the following two matrices respectively:

a11a12a13a14a15a21a22a23a24a25a31a32a33a32a31a25a24a23a22a21a15a14a13a12a11anda11a12a13a14a15a16a21a22a23a24a25a26a31a32a33a34a35a36a36a35a34a33a32a31a26a25a24a23a22a21a16a15a14a13a12a11.

The class of CS matrices plays an important role in the study of Markov processes since they are indeed transition matrices for some processes in evolutionary biology. For instance, in Kimura (1957), centrosymmetric matrices are used to study the random assortment phenomena of subunits in chromosome division. Furthermore, in Schensted (1958), the same centrosymmetric matrices appear as the transition matrices in the model of subnuclear segregation in the macronucleus of ciliates. Finally, the work (Iosifescu 2014) examines a special case of the random genetic drift phenomenon, which consists of a population of individuals that are able to produce a single type of gamete. In this case, the transition matrices of the associated Markov chain are given by centrosymmetric matrices.

The embedding problem is directly related to the notions of matrix exponential and logarithm which we introduce for completeness below.

Definition 3

We define the exponential exp(A) of a matrix A, using the Taylor power series of the function f(x)=ex, as

exp(A)=k=0Akk!,

where A0=In and In denotes the n×n identity matrix. If A=Pdiag(λ1,,λn)P-1 is an eigendecomposition of A, then exp(A)=Pdiag(eλ1,,eλn)P-1. Given a matrix AMn(K), a matrix BMn(K) is said to be a logarithm of A if exp(B)=A. If v is an eigenvector corresponding to the eigenvalue λ of A, then v is an eigenvector corresponding to the eigenvalue eλ of exp(A).

A Markov matrix M is called embeddable if it can be written as the exponential of a rate matrix Q, namely M=exp(Q). Then any rate matrix Q satisfying the equation M=exp(Q) is called a Markov generator of M.

Remark 1

Embeddable Markov matrices occur when we assume a continuous time Markov chain, in which case the Markov matrices have the form

M=exp(tQ),

where t0 represents time and Q is a rate matrix. However, in the rest of the paper, we assume that t is incorporated in the rate matrix Q.

The existence of multiple logarithms is a direct consequence of the distinct branches of the logarithmic function in the complex field.

Definition 4

Given zC\R0 and kZ, the k-th branch of the logarithm of z is logk(z):=log|z|+(Arg(z)+2πk)i, where log is the logarithmic function on the real field and Arg(z)(-π,π) denotes the principal argument of z. The logarithmic function arising from the branch log0(z) is called the principal logarithm of z and is denoted as log(z).

It is known that if A is a matrix with no negative eigenvalues, then there is a unique logarithm of A all of whose eigenvalues are given by the principal logarithm of the eigenvalues of A (Higham 2008, Theorem 1.31). We refer to this unique logarithm as the principal logarithm of A, denoted by Log(A).

By definition, the Markov generators of a Markov matrix M are those logarithms of M that are rate matrices. In particular they are real logarithms of M. The following result enumerates all the real logarithms with rows summing to zero of any given Markov matrix with positive determinant and distinct eigenvalues. Therefore, all Markov generators of such a matrix are necessarily of this form.

Proposition 1

(Casanellas et al. 2023, Proposition 4.3). Let M=Pdiag(1,λ1,,λt,μ1,μ1¯,,μs,μs¯)P-1 be an n×n Markov matrix with PGLn(C) and distinct eigenvalues λiR>0 for i=1,,t and μj{zC:Im(z)>0} for j=1,,s, all of them pairwise distinct. Then, a matrix Q is a real logarithm of M with rows summing to zero if and only if Q=Pdiag(0,log(λ1),,log(λt),logk1(μ1),logk1(μ1)¯,,logks(μs),logks(μs)¯)P-1 for some k1,,kjZ.

Remark 2

In particular, the principal logarithm of M can be computed as

Log(M)=Pdiag(0,log(λ1),,log(λt),log(μ1),log(μ1¯),,log(μs),log(μs¯))P-1.

In this paper, we focus on the embedding problem for the class of centrosymmetric matrices. In Sect. 3, we will first study the embeddability of 4×4 centrosymmetric Markov matrices, which include the K3P, K2P and JC Markov matrices. In Sect. 5 and Sect. 6, we will further study the embeddability of higher order centrosymmetric Markov matrices.

Embeddability of 4×4 centrosymmetric matrices

In this section, we begin our study by analyzing the embeddability of 4×4 centrosymmetric matrices also known as strand symmetric matrices. We will provide necessary and sufficient conditions for 4×4 centrosymmetric matrices to be embeddable. Moreover, we will discuss their rate identifiability problem as well.

The transition matrices of 4×4 centrosymmetric matrices are assumed to have the form

M=m11m12m13m14m21m22m23m24m24m23m22m21m14m13m12m11,

where

m11+m12+m13+m14=1=m21+m22+m23+m24andmij0.

Recall that the K3P matrices are assumed to have the form

M=m11m12m13m14m12m11m14m13m13m14m11m12m14m13m12m11.

In the case of the K2P matrices, we additionally have m12=m13, while in the case of JC matrices, m12=m13=m14. It can be easily seen that K3P, K2P, and JC Markov (rate) matrices are centrosymmetric.

Let us define the following matrix

S=1001011001-10100-1; 3.1

compare (Casanellas and Kedzierska 2013, Section 6). For a 4×4 CS Markov matrix M, we define F(M):=S-1MS. By direct computation, it can be checked that F(M) is a block diagonal matrix

F(M)=λ1-λ001-μμ0000αα00ββ, 3.2

where

λ=m11+m14,μ=m22+m23,α=m22-m23,α=m21-m24,β=m11-m14,β=m12-m13. 3.3

Define two matrices, M1:=λ1-λ1-μμ and M2:=ααββ, which are the upper and lower block matrices in (3.2), respectively.

Similarly, the rate matrices in strand symmetric models are assumed to have the 4×4 centrosymmetric form

Q=q11q12q13q14q21q22q23q24q24q23q22q21q14q13q12q11,

where

q11+q12+q13+q14=0=q21+q22+q23+q24andqij0forij.

So, for a 4×4 CS rate matrix Q, we can also define F(Q):=S-1QS. By direct computation, it can be checked that

F(Q)=-ρρ00σ-σ0000δδ00γγ, 3.4

where

ρ=q12+q13,σ=q21+q24,δ=q22-q23,δ=q21-q24,γ=q11-q14,γ=q12-q13.

Define two matrices, Q1:=-ρρσ-σ and Q2:=δδγγ, which are the upper and lower block matrices in (3.4), respectively.

The following results provide necessary conditions for a 4×4 CS Markov matrix to be embeddable.

Lemma 1

Let M=(mij) be a 4×4 CS Markov matrix and M=exp(Q) for some CS rate matrix Q. Then

  1. m11+m14+m22+m23>1 and

  2. (m22-m23)(m11-m14)>(m24-m21)(m13-m12).

Proof

We have that

F(M)=S-1MS=S-1exp(Q)S=exp(S-1QS)=exp(F(Q)).

Then

M100M2=exp(F(Q))=exp(Q1)00exp(Q2).

Thus, M1 is an embeddable 2×2 Markov matrix. Using the embeddability criteria of 2×2 Markov matrices in Kingman (1962), we have that 1<tr(M1)=λ+μ, which is the desired inequality. Additionally, since M2=exp(Q2), det(M2)>0 as desired.

Lemma 2

Let M=(mij) be a 4×4 CS Markov matrix and M=exp(Q) for some CS rate matrix Q=(qij). If λ+μ2, then

q12+q13=-λ+1λ+μ-2ln(λ+μ-1)

and

q21+q24=-μ+1λ+μ-2ln(λ+μ-1).

Proof

By direct computations and the proof of Lemma 1,

M1=exp(Q1)=1ρ+σe-ρ-σρ+σ-e-ρ-σρ+ρ-e-ρ-σσ+σe-ρ-σσ+ρ.

We then have the following system of equations:

λ=e-ρ-σρ+σρ+σandμ=e-ρ-σσ+ρρ+σ. 3.5

Summing the two equations, we get

λ+μ=e-ρ-σ+1.

Note that by Lemma 1, λ+μ>1. Therefore,

ρ+σ=-ln(λ+μ-1). 3.6

Using Equation (3.5) and (3.6), we obtain

ρ=-λ+1λ+μ-2ln(λ+μ-1)andσ=-μ+1λ+μ-2ln(λ+μ-1).

The proof is now complete.

Proposition 2

Given two matrices A=(aij),B=(bij)M2(R), consider the block-diagonal matrix C=diag(A,B). Then the following statements hold:

  • i)

    F-1(C):=SCS-1 is a CS matrix.

  • ii)
    F-1(C) is a Markov matrix if and only if A is a Markov matrix and
    |b22|a11,|b21|a12,|b12|a21,|b11|a22.
  • iii)
    F-1(C) is a rate matrix if and only if A is a rate matrix and
    b22a11(0),|b21|a12(=-a11),|b12|a21(=-a22),b11a22(0).

Proof

To prove i), by direct computation we obtain that

F-1(C)=SCS-1=12a11+b22a12+b21a12-b21a11-b22a21+b12a22+b11a22-b11a21-b12a21-b12a22-b11a22+b11a21+b12a11-b22a12-b21a12+b21a11+b22.

Then ii) follows from the above expression of F-1(Q) and the fact that rows of Markov matrices add to 1 and the entries are non-negative, while iii) similarly follows from the fact that the rows of rate matrices add to zero and the off-diagonal entries are non-negative.

For any 4×4 CS Markov matrix M=(mij), let us recall that by (3.2), M is block-diagonalizable via the matrix S. In the rest of this section, we will study both the upper and the lower block matrices of F(M) more closely. Studying the upper and lower blocks allows us to establish the main result of the embeddability criteria for 4×4 CS Markov matrices. This block-diagonalization reduces our analysis to studying the logarithms of both the upper and the lower block matrices which have size 2×2. This result will be presented in Theorem 2.

Upper block

As we have seen in (3.2), the upper block of F(M) is given by the 2×2 matrix M1=λ1-λ1-μμ, which is a Markov matrix. If P1=11-λ1μ-1, then

P1-1M1P1=100λ+μ-1.

Hence, by Proposition 1, any logarithm of M1 can be written as

Lk1,k2M1:=P12k1πi00log(λ+μ-1)+2k2πiP1-1,

for some integers k1 and k2. Let p=log(λ+μ-1), q=1-λ, and r=1-μ. Then

Lk1,k2M1=12-λ-μqp+2π(rk1+qk2)i-qp+2πq(k1-k2)i-rp+2πr(k1-k2)irp+2π(qk1+rk2)i. 3.7

Lemma 3

If λ+μ2, then Lk1,k2M1 is a real matrix if and only if k1=k2=0 and λ+μ>1. In this case, the only real logarithm of M1 is the principal logarithm

12-λ-μqp-qp-rprp.

Proof

For fixed k1 and k2, the eigenvalues of Lk1,k2M1 are λ1=2k1πi and λ2=p+2k2πi. Then Lk1,k2M1 is a real matrix if and only if λ1,λ2R or λ2=λ1¯. Since λ+μ2, λ2λ1¯. Thus, Lk1,k2M1 is a real matrix if and only if λ1,λ2R. Finally, λ1R if and only if k1=0 and λ2R if and only if k2=0 and λ+μ>1.

Lower block

The lower block of F(M) is given by the matrix M2=ααββ. Unlike M1, the matrix M2 is generally not a Markov matrix. The discriminant of the characteristic polynomial of M2 is given by

Δ:=(α-β)2+4αβ 3.8

with α,β,α,β defined as in (3.3). If Δ>0, then M2 has two distinct real eigenvalues and if Δ<0, then M2 has a pair of conjugated complex eigenvalues. Moreover, if Δ=0, then M2 has either 2×2 Jordan block or a repeated real eigenvalue. We will assume that Δ0 so that M2 diagonalizes into two distinct eigenvalues.

Let P2=Δ+(α-β)2Δ-(α-β)2β-β. Then

P2-1M2P2=(α+β)+Δ200(α+β)-Δ2.

Let us now define

l3:=log((α+β)+Δ2)+2k3πiandl4:=log((α+β)-Δ2)+2k4πi,

where k3 and k4 are integers. Therefore, any logarithm of M2 can be written as

Lk3,k4M2:=εϕγη 3.9

where

ε:=12((l3+l4)+(α-β)(l3-l4)Δ),ϕ:=α(l3-l4)Δ,γ:=β(l3-l4)Δandη:=12((l3+l4)-(α-β)(l3-l4)Δ).

Lemma 4

  1. If Δ>0, then Lk3,k4M2 is a real matrix if and only if α+β>Δ and k3=k4=0.

  2. If Δ<0, then Lk3,k4M2 is a real matrix if and only if k4=-k3.

Proof

  1. If Δ>0, then Im(l3)=2k3π and Im(l4)=2k4π. Moreover, Re(l3)Re(l4). Since l3 and l4 are the eigenvalues of Lk3,k4M2, this implies that l3l¯4. In particular, Lk3,k4M2 is a real matrix if and only if both l3 and l4 are real.

  2. Let us assume Δ<0 and take z=(α+β)+Δ2. Fixing k3,k4Z, the eigenvalues of Lk3,k4M2 are l3=log(z)+2k3πi and l4=log(z¯)+2k4πi=log(z)¯+2k4πi, which are both complex numbers. Thus, Lk3,k4M2 is real if and only if l3=l¯4. Hence, k4=-k3. Conversely, k4=-k3 implies that l3+l4=2Re(l3)R and l3-l4Δ=2Im(l3)iΔR. Thus, all entries of Lk3,k4M2 are real.

Logarithms of 4×4 CS Markov matrices

Let M be a 4×4 CS Markov matrix. Using the values defined in (3.3) and (3.8), we can now label up its four eigenvalues as following,

1,λ1:=λ+μ-1,λ2:=(α+β)+Δ2andλ3=(α+β)-Δ2. 3.10

We note that the subset of 4×4 CS Markov matrix with repeated eigenvalues (diagonalizing matrix with repeated eigenvalues or a Jordan block of size greater than 1) have zero measure. Therefore generic 4×4 Markov matrices have no repeated eigenvalues, and hence we are going to assume the eigenvalues to be distinct. In particular, we are assuming that M diagonalizes. Furthermore, since we want M to have real logarithms and have no repeated eigenvalues, we need the real eigenvalues to be positive.

The following theorem characterizes the embeddability of a 4×4 CS Markov matrix with positive and distinct eigenvalues. Furthermore, the theorem guarantees that a 4×4 CS Markov matrix is embeddable if and only if it admits a CS Markov generator. In particular, the characterization of the embeddability of a CS matrix is equivalent when restricting to rate matrices satisfying the symmetries imposed by the model (model embeddability) than when restricting to all possible rate matrices (embedding problem).

Theorem 1

Let M be a diagonalizable 4×4 CS Markov matrix with positive and distinct eigenvalues λ1,λ2,λ3 defined as in (3.10). Let us define

x=log(λ1),yk=log(λ2)+2kπi,zk=log(λ3)-2kπi,

where k=0 if Δ>0 and kZ if Δ<0. Then any real logarithm of M is given by

Sα1-α100-β1β10000δ(k)ε(k)00ϕ(k)γ(k)S-1,

where

α1=1-λ2-λ-μx,β1=1-μ2-λ-μx,δ(k)=12((yk+zk)+(α-β)(yk-zk)Δ),ε(k)=α(yk-zk)Δ,ϕ(k)=β(yk-zk)Δ,γ(k)=12((yk+zk)-(α-β)(yk-zk)Δ).

with λ,μ,α,β,αandβ defined as in (3.3) and Δ as in (3.8).

In particular, any real logarithm of M is also a 4×4 CS matrix whose entries q11,,q24 are given by:

q11=α1+γ(k)2,q12=-α1+ϕ(k)2,q13=-α1-ϕ(k)2,q14=α1-γ(k)2,q21=-β1+ε(k)2,q22=β1+δ(k)2,q23=β1-δ(k)2,q24=-β1-ε(k)2.

Proof

Let us note that

M=S·diag(P1,P2)·diag(1,λ1,λ2,λ3)·diag(P1-1,P2-1)·S-1.

Since we assume that the eigenvalues of M are distinct, according to Proposition 1, any logarithm of M can be written as

Q=S·diag(P1,P2)·diag(logk1(1),logk2(λ1),logk3(λ2),logk4(λ3))·diag(P1-1,P2-1)·S-1=S·diag(Lk1,k2M1,Lk3,k4M2)·S-1,

The last equation and the fact that S and S-1 are real matrices imply that Q will be real if and only if both Lk1,k2M1 and Lk3,k4M2 are real. Here Lk1,k2M1 is the upper block given in (3.7) and Lk3,k4M2 is the lower block defined in (3.9). By Lemma 3, Lk1,k2M1 being a real logarithm implies that k1=k2=0 and λ+μ>1. Then Lk3,k4M2 being a real matrix, according to Lemma 4, implies that k3=k4=0 if Δ>0, while k4=-k3 if Δ<0. Therefore, the upper block is L0,0M1 and the lower block will be Lk,-kM2,fork=k3 completing the proof.

Now we are interested in knowing when the real logarithm of a 4×4 CS Markov matrix is a rate matrix. Using the same notation as in Theorem 1 we get the following result.

Theorem 2

A diagonalizable 4×4 CS Markov matrix M with distinct eigenvalues is embeddable if and only if the following conditions hold for k=0 if Δ>0 or for some kZ if Δ<0:

λ1>0,(α+β)2>Δ,|ϕ(k)|-α1,|ε(k)|-β1,γ(k)α1,δ(k)β1.

Proof

The logarithm of a 4×4 CS Markov matrix will depend on whether Δ>0 or Δ<0. In particular, it will depend on whether the eigenvalues λ2 and λ3 are real and positive or whether they are conjugated complex numbers.

  1. If Δ>0, then both λ2 and λ3 are real and λ2>λ3. Hence, z<y<0. Moreover, Lemma 4 implies that λ3>0 and hence λ2λ3>0.

  2. If Δ<0, then λ2,λ3C\R and λ2=λ3¯. Hence, y+z>0 and y-z=4πki. Moreover, λ2λ3=|λ3|2>0 since λ30.

Thus, in both cases, α1,β1,δ(k),ε(k),ϕ(k),γ(k)R. Moreover, α1 and β1 are both non-positive. In particular, Theorem 1 together with Proposition 2 imply that a real logarithm of M is a rate matrix if and only if

|ϕ(k)|-α1,|ε(k)|-β1,γ(k)α1,δ(k)β1.

Furthermore, the conditions λ1>0 comes from Lemma 3. The proof is now complete.

Remark 3

According to Theorem 2 the embeddability of a 4×4 CS Markov matrix M with distinct positive eigenvalues can be decided by checking six inequalities depending on the entries of M. However, if M has non-real eigenvalues then one has to check infinitely many groups of inequalities, one for each value of kZ. It is enough that one of those systems is consistent to guarantee that M is embeddable. Theorem 5.5 in Casanellas et al. (2023) provides boundaries for the values of k for which the corresponding inequalities may hold.

Let us take a look at the class of K3P matrices which is a special case of strand symmetric matrices. Indeed, for a K3P matrix M=(mij), we have that

m11=m22,m12=m21,m13=m24andm14=m23.

Suppose that a K3P-Markov matrix M=(mij) is K3P-embeddable, i.e. M=exp(Q) for some K3P-rate matrix Q. Recall that the eigenvalues of M are

1,p:=m11+m12-m13-m14,q:=m11-m12+m13-m14andr:=m11-m12-m13+m14.

In this case, we have that

λ=μ=m11+m14,α=β=m11-m14,α=β=m13-m12,λ1=randΔ=4(m13-m12)2.

In particular, we see that Δ>0 unless m12=m13. Moreover,

x=logr,y=logq,z=logp,α1=β1=12logr,δ(0)=γ(0)=12logpq,|ε(0)|=|ϕ(0)|=12logqp.

The inequalities in Theorem 2 can be spelled out as follows:

r>0,pq>0,|logqp|-lograndlogpqlogr.

These inequalities are equivalent to the K3P-embeddability criteria presented in (Roca-Lacostena and Fernández-Sánchez 2018b, Theorem 3.1) and (Ardiyansyah et al. 2021, Theorem 1). Moreover, they are also equivalent to the restriction to centrosymmetric-matrices of the embeddability criteria for 4×4 Markov matrices with different eigenvalues given in (Casanellas et al. 2023, Theorem 1.1)

In the last part of this section, we discuss the rate identifiability problem for 4×4 centrosymmetric matrices. If a centrosymmetric Markov matrix arises from a continuous-time model, then we want to determine its corresponding substitution rates. In other words, given an embeddable 4×4 CS matrix, we want to know if we can uniquely identify its Markov generator.

It is worth noting that Markov matrices with repeated real eigenvalues may admit more than one Markov generator (e.g. examples 4.2 and 4.3 in (Casanellas et al. 2020a) show embeddable K2P matrices with more than one Markov generator). Nonetheless, this is not possible if the Markov matrix has distinct eigenvalues, because in this case its only possible real logarithm would be the principal logarithm (Culver 1966). As one considers less restrictions in a model, the measure of the set of matrices with repeated real eigenvalues decreases, eventually becoming a measure zero set. For example, this is the case within the K3P model, where both its submodels (the K2P model and the JC model) consist of matrices with repeated eigenvalues and have positive measure subsets of embeddable matrices with non-identifiable rates. However, when considering the whole set of K3P Markov matrices, the subset of embeddable matrices with more than one Markov generator has measure zero (see Chapter 4 in (Roca-Lacostena 2021)). Nevertheless, this behaviour only holds if the Markov matrices within the model have real eigenvalues.

Proposition 3

There is a positive measure subset of 4×4 CS Markov matrices that are embeddable and whose rates are not identifiable. Moreover, all the Markov generators of the matrices in this set are also CS matrices.

Proof

Given

P=1-51-i1+i12-ii12i-i1-5-1+i-1-i,

let us consider the following matrices

M=Pdiag(1,e-7π,e-4πi,-e-4πi)P-1,Q=Pdiag(0,-7π,-4π-3π2i,-4π+3π2i)P-1.

A straightforward computation shows that M is a CS Markov matrix and Q is a CS rate matrix. Moreover they both have non-zero entries. By applying the exponential series to Q, we get that exp(Q)=M. This means that M is embeddable and Q is a Markov generator of M.

Since Q is a rate matrix, so is Qt for any tR0. Therefore, exp(Qt) is an embeddable Markov matrix, because the exponential of any rate matrix is necessarily a Markov matrix. See (Pachter and Sturmfels 2005, Theorem 4.19) for more details. Moreover, we have that

S-1P=1-500120000-ii001-i1+i,

so S-1exp(Qt)S is a 2-block diagonal matrix. Hence, by Proposition 2 we have that exp(Qt) is an embeddable strand symmetric Markov matrix for all tR>0.

Now, let us define V=Pdiag(0,0,2πi,-2πi)P-1. Note that Q and V diagonalize simultaneously via P and hence they commute. Therefore,

exp(Q+V)=exp(Q)exp(V)=MI4=M

by the Baker-Campbell-Haussdorff formula. Moreover,

exp(Qt+kV)=exp(Qt)exp(kV)=exp(Qt)I4=exp(Qt)

for all kZ. Note that kV is a bounded matrix for any given k and hence, given t large enough, it holds that Qt+mV is a rate matrix for any m between 0 and k.

This shows that, for t large enough, exp(Qt) is an embeddable CS Markov matrix with at least k+1 different CS Markov generators. Moreover, exp(Qt) and all its generators have no null entries by construction and they can therefore be perturbed as in Theorem 3.3 in Casanellas et al. (2020b) to obtain a positive measure subset of embeddable CS Markov matrices that have k+1 CS Markov generators.

Remark 4

The perturbation presented in Theorem 3.3 in Casanellas et al. (2020b) consists of small changes on the real and complex parts of the eigenvalues and eigenvectors of M other than the eigenvector (1,,1) and its corresponding eigenvalue 1. If those changes are small enough then the resulting transition matrix and all its generators still satisfy the stochastic constraints of Markov/rate matrices.

Remark 5

Using the same notation as in the proposition above and given CGL2(C), let us define

Q(C)=Pdiag(I2,C)diag1,-7π,-4π-3π2i,-4π+3π2idiag(I2,C-1)P-1.

Since Q(I2)=Q is a CS rate matrix with no null entries, so is Q(C) for CGL2(C) close enough to I2. Moreover, by construction we have that exp(2tQ(C))=exp(2tQ) for all tN. Therefore, for tN we have that exp(2tQ) has uncountably many Markov generators (i.e. 2tQ(C) with C close to I2) and all of them are CS matrices (Culver 1966, Corollary 1). It is worth noting that according to (Culver 1966, Corollary 1), if a matrix has uncountably many logarithms, then it necessarily has repeated real eigenvalues. Therefore, the subset of embeddable CS Markov matrices with uncountably many generators has measure zero within the set of all matrices.

Volumes of 4×4 CS Markov matrices

In this section, we compute the relative volumes of embeddable 4×4 CS Markov matrices within some meaningful subsets of Markov matrices. The aim of this section is to describe how large the different sets of matrices are compared to each other.

Let V4Markov be the set of all 4×4 CS Markov matrices. We use the following description

V4Markov={(b,c,d,e,g,h)TR6:b,c,d,e,g,h0,1-b-c-d0,1-e-g-h0}.

More explicitly, we identify the 4×4 CS Markov matrix

1-b-c-dbcde1-e-g-hghhg1-e-g-hedcb1-b-c-d

with a point (b,c,d,e,g,h)V4Markov. Let V+ be the set of all CS Markov matrices having real positive eigenvalues, where

Δ=((1-e-2g-h)-(1-b-c-2d))2+4(e-h)(b-c),

is the discriminant of the matrix M2 as stated in Sect. 3. We have V+V4Markov. More explicitly,

V+={(b,c,d,e,g,h)R6:b,c,d,e,g,h0,1-b-c-d0,1-e-g-h0,1-b-c-e-h>0,(2-b-c-2d-e-2g-h)+Δ>0,(2-b-c-2d-e-2g-h)-Δ>0,Δ>0}.

Let Vem+ be the set of all embeddable 4×4 CS Markov matrices with four distinct real positive eigenvalues. We have Vem+V+. Therefore, by Theorem 2,

Vem+={(b,c,d,e,g,h)R6:b,c,d,e,g,h0,1-b-c-d0,1-e-g-h0,1-b-c-e-h>0,(2-b-c-2d-e-2g-h)+Δ>0,(2-b-c-2d-e-2g-h)-Δ>0,Δ>0,|ϕ(0)|-α1,|ε(0)|-β1,δ(0)β1,γ(0)α1}.

Finally, we consider the following two biologically relevant subsets of V4Markov. Let VDLC be the set of diagonally largest in column (DLC) Markov matrices, which is the subset of V4Markov containing all CS Markov matrices such that the diagonal element is the largest element in each column. These matrices are related to matrix parameter identifiability in phylogenetics (Chang 1996). Secondly, we let VDD be the set of diagonally dominant (DD) Markov matrices, which is the subset of V4Markov matrices containing all CS Markov matrices such that in each row the diagonal element is at least the sum of all the other elements in the row. Biologically, the subspace VDD consists of matrices with probability of not mutating at least as large as the probability of mutating. If a diagonally dominant matrix is embeddable, it has an identifiable rate matrix (Cuthbert 1972; James 1973). By the definition of each set, we have the inclusion VDDVDLC.

Remark 6

The sets V+, Vem+, VDLC, VDD that we consider in this section are all subsets of the set V4Markov of all 4×4 CS Markov matrices, but we can use the same definition to refer to the equivalent subsets of n×n CS Markov matrices. Therefore, we will use the same notation V+, Vem+, VDLC, VDD to refer to the equivalent subsets of the set VnMarkov of n×n CS Markov matrices without confusion in the following sections.

In the rest of this section, the number v(A) denotes the Euclidean volume of the set A. By definition, V4Markov, VDLC and VDD are polytopes, since they are defined by the linear inequalities in R6. Hence, we can use Polymake (Gawrilow and Joswig 2000) to compute their exact volumes and obtain that

v(V4Markov)=136,v(VDLC)=1576andv(VDD)=12304.

Hence, we see that VDLC and VDD constitute roughly only 6.25% and 1.56% of V4Markov, respectively.

On the other hand, we will estimate the volume of the sets V+,Vem+,VDLCV+,VDLCVem+,VDDV+,andVDDVem+ using the hit-and-miss Monte Carlo integration method (Hammersley 2013) with sufficiently many sample points in Mathematica (Inc. 2022). Theoretically, Theorem 2 enables us to compute the exact volume of these relevant sets. For example in the case of K3P matrices, such exact computation of volumes has been feasible in Roca-Lacostena and Fernández-Sánchez (2018b). However, while for the K3P matrices, the embeddability criterion is given by three quadratic polynomial inequalities, in the case of CS matrices the presence of nonlinear and nonpolynomial constraints imposed on each set, makes the exact computation of the volume of these sets intractable. Therefore, we need to approximate the volume of these sets. Given a subset AV4Markov, the volume estimate of v(A) computed using the hit-and-miss Monte Carlo integration method with n sample points is given by the number of points belonging to A out of n sample points. For computational purposes, in the formula of ϕ(0) and ε(0), we use the fact that

y-z=log(2-b-c-2d-e-2g-h)+Δ(2-b-c-2d-e-2g-h)-Δ.=log((2-b-c-2d-e-2g-h)+Δ)2(2-b-c-2d-e-2g-h)2-Δ.=log((2-b-c-2d-e-2g-h)+(b+c+2d-e-2g-h)2+4(e-h)(b-c))2(2-b-c-2d-e-2g-h)2-((b+c+2d-e-2g-h)2+4(e-h)(b-c))

All codes for the computations implemented Mathematica and Polymake can be found at the following address: https://github.com/ardiyam1/Embeddability-and-rate-identifiability-of-centrosymmetric-matrices.

The results of these estimations using the hit-and-miss Monte Carlo integration implemented in Mathematica with n sample points are presented in Table 1, while Table 2 provides an estimated volume ratio between relevant subsets of centrosymmetric Markov matrices using again the hit-and-miss Monte Carlo integration with n sample points. In Table 1, we firstly generate n centrosymmetric matrices whose off-diagonal entries were sampled uniformly in [0, 1] and forced the rows of the matrix to sum to one. Out of these n matrices, we test how many of them are actually Markov matrices (i.e. the diagonal entries are non-negative) and then out of these how many have positive eigenvalues. In particular, for n=107 sample points containing 277628 centrosymmetric Markov matrices, Table 2 suggests that there are approximately 1.7% of centrosymmetric Markov matrices with distinct positive eigenvalues that are embeddable. Moreover, we can see that for n=107, out of all embeddable centrosymmetric Markov matrices with distinct positive eigenvalues, almost all are diagonally largest in column, while only 28% are diagonally dominant.

Table 1.

Number of samples in the sets

n 104 105 106 107
Samples in V4Markov 280 2767 27829 277628
Samples V+ 23 192 1999 20601
Samples in Vem+ 3 34 359 3511
Samples in VDLCV+ 19 154 1541 15830
Samples in VDDV+ 3 31 262 2889
Samples in VDLCVem+ 3 34 357 3503
Samples in VDDVem+ 1 15 105 1011

V+,Vem+,VDLCV+,VDLCVem+,VDDV+andVDDVem+ using hit-and-miss methods and Theorem 2.

Table 2.

Relative volumes ratio between the relevant subsets obtained using hit-and-miss method and Theorem 2. The volumes were estimated as the quotient of the sample sizes in Table 1

n 104 105 106 107
v(Vem+)v(V+) 0.130435 0.177083 0.17959 0.170429
v(VDLCVem+)v(VDLCV+) 0.157895 0.220779 0.231668 0.221289
v(VDLCVem+)v(V+) 0.130435 0.177083 0.178589 0.17004
v(VDLCVem+)v(Vem+) 1 1 0.994429 0.997721
v(VDDVem+)v(VDDV+) 0.333333 0.483871 0.400763 0.349948
v(VDDVem+)v(V+) 0.0434783 0.078125 0.052563 0.0490753
v(VDDVem+)v(Vem+) 0.333333 0.441176 0.292479 0.287952

An alternative approach for approximating the number of embeddable matrices within the model is to use Algorithm 5.8 in Casanellas et al. (2023) to test the embeddability of the sample points. Tables  4 and 5 below are analogous to Tables 1 and 2, but Table 4 was obtained using the sampling method in (Roca-Lacostena 2021, Appendix A), while using either Algorithm 5.8 in Casanellas et al. (2023) or the inequalities in Theorem 2 yields identical results which are provided in Tables 4 and 5.

Table 4.

Number of samples in Vem+,VDLCVem+andVDDVem+ obtained by applying either Theorem 2 or the results in Casanellas et al. (2023) on the sample set in Table 3

Samples in Vem+ 1877 18663 185357 1862413
Samples in VDLCVem+ 1869 18586 184555 1854592
Samples in VDDVem+ 516 5164 50058 504304

Table 5.

Relative volumes ratio between the relevant subsets obtained using hit-and-miss method and either Algorithm 5.8 in Casanellas et al. (2023) or Theorem 2. The volumes were estimated as the quotient of the sample sizes in Tables 3 and 4

n 104 105 106 107
v(Vem+)v(V+) 0.1877 0.18663 0.185357 0.1862413
v(VDLCVem+)v(VDLCV+) 0.2191 0.2175 0.2159 0.2169
v(VDLCVem+)v(V+) 0.1869 0.18586 0.184555 0.1854592
v(VDLCVem+)v(Vem+) 0.9957 0.9959 0.99567 0.99580
v(VDDVem+)v(VDDV+) 0.3524 0.3552 0.3463 0.3481
v(VDDVem+)v(V+) 0.0516 0.05164 0.050058 0.0504
v(VDDVem+)v(Vem+) 0.2749 0.2767 0.2701 0.2708

We used the python implementation of Algorithm 5.8 in Casanellas et al. (2023) provided in (Roca-Lacostena 2021, Appendix A) and modified it to sample on the set of 4×4 CS Markov matrices with positive eigenvalues. The original sampling method used in (Roca-Lacostena 2021, Appendix A) consisted of sampling uniformly on the set of 4×4 centrosymmetric-Markov matrices until we obtained n samples (or as many samples as we require) with positive eigenvalues.

Despite the fact that Theorem 2 and Algorithm 5.8 in Casanellas et al. (2023) were originally implemented using different programming languages (Wolfram Mathematica and Python respectively) and were tested with different sample sets, the results obtained are quite similar as illustrated by Tables 2 and 5. In fact, when we apply both Algorithm 5.8 in Casanellas et al. (2023) and Theorem 2 on the same sample set in Table 3, we obtain identical results which are displayed in Tables 4 and 5.

Table 3.

Number of samples in V+,VDLCV+,andVDDV+ obtained by using the sampling method in (Roca-Lacostena 2021, Appendix A)

Samples in V+ 104 105 106 107
Samples in VDLCV+ 8531 85446 854709 8549100
Samples in VDDV+ 1464 14538 144546 1448720

It is worth noting that the embeddability criteria given in Theorem 2 use inequalities depending on the entries of the matrix, whereas Algorithm 5.8 in Casanellas et al. (2023) relies on the computation of its principal logarithm and its eigenvalues and eigenvector, which may cause numerical issues when working with matrices with determinant close to 0. Moreover, the computation of logarithms can be computationally expensive. As a consequence, the algorithm implementing the criterion for embeddability arising from Theorem 2 is faster. Table 6 shows the running times for the implementation of both embeddability criteria used to obtain Table 5.

Table 6.

Running times for the Python implementation of the embeddability criterion arising from Theorem 2 and from Algorithm 5.8 in Casanellas et al. (2023). The simulations were run using a computer with 8GB of memory

104 105 106 107
Sampling time 12.5s 121.5s (2 min) 1222s (20min) 12141.8s (3h 22min)
Embedding criteria (Theorem 2) 28.3s 273.2s (4min 30s) 2703s (45min) 27413s (7h 37min)
Embedding criteria (Algorithm 5.8) 84.2s 840.5s (15 min) 8358 (2h 19min) 83786s (23h 16min)

The Python implementation of Algorithm 5.8 in Casanellas et al. (2023) provided in (Roca-Lacostena 2021, Appendix A) can also be used to test the embeddability of any 4×4 CS Markov matrix (including those with non-real eigenvalues) without modifying the embeddability criteria. To do so, it is enough to apply the algorithm to a set of Markov matrices with different eigenvalues sampled uniformly from the set of all 4×4 CS Markov matrix. As hinted in Remark 3, this would also be possible using the embedability criterion in Theorem 2 together with the boundaries for k provided in (Casanellas et al. 2023, Theorem 5.5). Table 7 shows the results obtained when applying Algorithm 5.8 in Casanellas et al. (2023) to a set of 107 4×4 CS Markov matrices sampled uniformly.

Table 7.

Embeddable matrices within 4×4 CS Markov matrices and its intersection with DLC matrices and DD matrices

Samples Embeddable samples Proportion of embeddable
V4Markov 107 173455 0.0173455
VDLC 1021195 172380 0.1688022
VDD 156637 49471 0.3158321

As most DLC and DD matrices have positive eigenvalues, the proportion of embeddable matrices within these subsets is almost the same when admitting matrices with non-positive eigenvalues (as in Table 7 instead of only considering matrices with positive eigenvalues as we did in Tables 2 and 5. On the other hand, the proportion of 4×4 embeddable CS matrices is much smaller in this case.

Centrosymmetric matrices and generalized Fourier transformation

In Sects. 3 and 4 we have seen the embeddability criteria for 4×4 centrosymmetric Markov matrices and the volume of their relevant subsets. In this section, we are extending this framework to larger matrices. The importance of this extension is relevant to the goal of synthetic biology which aims to expand the genetic alphabet. For several decades, scientists have been cultivating ways to create novel forms of life with basic biochemical components and properties far removed from anything found in nature. In particular, they are working to expand the number of amino acids which is only possible if they are able to expand the genetic alphabet (see for example (Hoshika et al. 2019)).

Properties of centrosymmetric matrices

For a fixed nN, let Vn denote the set of all centrosymmetric matrices of order n. Moreover, let VnMarkov and Vnrate denote the set of all centrosymmetric Markov and rate matrices of order n, respectively. As a subspace of the set of all n×n real matrices, for n even, dim(Vn)=n22 while for n odd, dim(Vn)=n2(n+1)+1. We will now mention some geometric properties of the sets VnMarkov and Vnrate. Furthermore, for any real number x, x and x denote the floor and the ceiling function of x, respectively.

Proposition 4

  1. For n even, VnMarkovR0n(n-1)2 is a Cartesian product of n2 standard (n-1)-simplices and its volume is 1(n-1)!n2. For n odd, VnMarkovR0n2n is a Cartesian product of n2 standard (n-1)-simplices and the n2-simplex with vertices {0,ei2}1in2{en2+1}, where ei is the i-th standard unit vector in Rn. Hence, the volume of VnMarkov is 12n2(n2)!(n-1)!n2.

  2. For n even, Vnrate=R0n(n-1)2 and for n odd, Vnrate=R0n2n.

Proof

Here we consider the following identification for an n×n centrosymmetric matrix M. For n even, M can be thought as a point (M1,,Mn2)(R0n)n2 where the point MiR0n corresponds to the i-th row of M. Similarly, for n odd, we identify M as a point in (R0n)n2×R0n2+1. Since M is a Markov matrix, under this identification, each point Mi lies in some simplices. Therefore, VnMarkov is a Cartesian product of some simplices. For n even, these simplices are the standard (n-1)-dimensional simplex:

x1++xn=1,xi0,1inx1++xn-11,xi0,1in-1 5.1

For n odd and 1in2, the point Mi belongs to standard (n-1)-simplex above and the point Mn2+1 belongs to the simplex

2x1++2xn2+xn2+1=1xi0,1in2+1x1++xn212,xi0,1in2 5.2

We now compute the volume of VnMarkov. Let us recall the fact that the volume of the Cartesian product of spaces is equal to the product of volumes of each factor space if the volume of each factor space is bounded. Moreover, the (n-1)-dimensional volume of the standard simplex in Eq. (5.1) in Rn-1 is 1(n-1)!. For n even, the statement follows immediately. For n odd, we use the fact that the n2-dimensional volume of the simplex in Eq. (5.2) is 12n2(n2)!. We refer the reader to Stein (1966) for an introductory text on the volume of simplices.

For the second statement, we use the fact that if Q is a rate matrix, then qii=-jiqij where qij0 for ij.

In the rest of this section, let Jn be the n×n anti-diagonal matrix, i.e. the (ij)-entries are one if i+j=n+1 and zero otherwise. The following proposition provides some properties of the matrix Jn that can be checked easily.

Proposition 5

Let A=(aij)Mn(R). Then

  1. (AJn)ij=ai,n+1-jand(JnA)ij=an+1-i,j.

  2. A is a centrosymmetric matrix if only if JnAJn=A.

In Sect. 3, we have seen that 4×4 CS matrices can be block-diagonalized through the matrix S. Now we will present a construction of generalized Fourier matrices to block-diagonalize any centrosymmetric matrices. Let us consider the following recursive construction of the n×n matrix Sn:

S1=1,S2=111-1andSn:=1010Sn-2010-1,forn3. 5.3

Proposition 6

For each natural number n3, Sn is invertible and its inverse is given by

Sn-1=120120Sn-2-10120-12.

Proof

The proposition easily follows from the definition of Sn. Indeed, we have

120120Sn-2-10120-12Sn=120120Sn-2-10120-121010Sn-2010-1=In.

The following proposition provides another block decomposition of the matrix Sn and its inverse.

Proposition 7

Let n2.

  1. For n even, Sn=In2Jn2Jn2-In2, while for n odd, Sn=In20Jn2010Jn20-In2.

  2. Using these block partitions, Sn-1=12Sn for n even, while Sn-1=12In2012Jn201012Jn20-12In2 for n odd.

Proof

The proof follows from induction on n and the fact that Jn2=In.

We will call a vector vRn symmetric if vi=vn+1-i for every 1in, i.e. Jnv=v. Moreover, we call a vector wRn anti-symmetric if vi=-vn+1-i for every 1in, i.e. Jnv=-v. The following technical proposition will be used in what follows in order to simplify a centrosymmetric matrix.

Proposition 8

Let n2. Let vRn be a symmetric vector and wRn be an anti-symmetric vector.

  1. The last n2 entries of Snv and vTSn are zero. Similarly, the last n2 entries of Sn-1v and vTSn-1 are zero.

  2. The first n2 entries of Snw and wTSn are zero. Similarly, the first n2 entries of Sn-1w and wTSn-1 are zero.

  3. Then the sum of the entries of Snv and vTSn is the sum of the entries of v.

  4. Then the sum of the entries of Sn-1v and vTSn-1 is the sum of the first n2 entries of v.

Proof

We will only prove the first part of item (1) in the proposition using mathematical induction on n. The base case for n=2 can be easily obtained. Suppose now that the proposition holds for all k<n. Let v=v1vv1Rn be a symmetric element. Then vRn-2 is also symmetric. By direct computation we obtain

Snv=1010Sn-2010-1v1vv1=2v1Sn-2v0.

The last n-22 entries of Sn-2v are zero. Thus, the last n-22+1=n2 entries of Snv are zero as well. The proof of the other statements can be obtained analogously using induction. In particular, let us note that the proof given for item (1) directly implies item (3).

For a fixed number n, let us define the following map:

Fn:Mn(R)Mn(R)AFn(A):=Sn-1ASn.

For n=4, we have seen that if A is a CS matrix, then F4(A) is a block-diagonal matrix where each block is of size 2×2 and is given by A1 and A2. Moreover, the upper block is a Markov matrix. The following lemma provides a generalization to these results.

Lemma 5

Let n2. Given an n×n CS matrix AFn(A) is the following block-diagonal matrix

Fn(A)=diag(A1,A2),

where A1 is a matrix of size n2×n2. Furthermore, if A is a Markov (rate) matrix, then A1 is also a Markov (rate) matrix.

Proof

First suppose that n is even. By (Cantoni and Butler 1976, Lemma 2), we can partition A into the following block matrices:

A=B1B2Jn2B2Jn2Jn2B1Jn2,

where B1 and B2 are of size n2×n2. By Proposition 7, we have

Sn-1ASn=12In2Jn2Jn2-In2B1B2Jn2B2Jn2Jn2B1Jn2In2Jn2Jn2-In2=B1+B2Jn200Jn2B1Jn2-Jn2B2.

Choose A1=B1+B2Jn2. Now suppose that A is a Markov matrix. This means that each row of A sums to 1 and A has non-negative entries. Therefore, for 1kn2, we have

j=1n2(A1)kj=j=1n2(B1+B2Jn2)kj=j=1n2(akj+ak,n2+j)=j=1nakj=1

and for 1jn2, (B1+B2Jn2)kj=akj+ak,n2+j0.

Now we consider the case when n is odd. Again by (Cantoni and Butler 1976, Lemma 2), we can partition A into the following block matrices:

A=B1pB2qrqJn2Jn2B2Jn2Jn2pJn2B1Jn2,

where B1,B2Mn2×n2(R), pandqM1×n2(R) and rM1×1(R). By Proposition 7, we have

Sn-1ASn=12In2012Jn201012Jn20-12In2B1pB2qrqJn2Jn2B2Jn2Jn2pJn2B1Jn2In20Jn2010Jn20-In2=B1+B2Jn2p02qr000Jn2B1Jn2-Jn2B2.

In this case, choose A1=B1+B2Jn2p2qr. Suppose that A is a Markov matrix. Since each row of A sums to 1, we have

j=1n22q1j+r=j=1nan2+1,j=1

and for 1kn2,

j=1n2(B1+B2Jn2)kj+pk1=j=1n2(akj+ak,n2+j+1)+ak,n2+1=j=1nakj=1.

From the fact that the entries of A are non-negative, for 1k,jn2, we obtain that

(B1+B2Jn2)kj=ak,j+ak,n2+j0.

Therefore, all entries of A1 sum to 1 and are non-negative meaning that A1 is a Markov matrix as well. We can proceed similarly for the case when A is a rate matrix.

Lemma 6

For any natural number n, let A1=(αi,j), A2=(βi,j)Mn2×n2(R). Suppose that Q=diag(A1,A2) is a block diagonal matrix. Then

  1. Fn-1(Q):=SnQSn-1 is a CS matrix.

  2. Fn-1(Q) is a Markov matrix if and only if A1 is a Markov matrix and for any 1i,jn2,
    αij+βn2+1-i,n2+1-j0andαi,n2+1-j-βn2+1-i,j0.
  3. Fn-1(Q) is a rate matrix if and only if A1 is a rate matrix and for any 1i,jn2, such that for i=j, αii+βn2+1-i,n2+1-i0 and for ij,
    αij+βn2+1-i,n2+1-j0andαi,n2+1-j-βn2+1-i,j0.

Proof

We will only prove the lemma for n even. Similar arguments will work for n odd as well. By Proposition 7,

Fn-1(Q)=12In2Jn2Jn2-In2A100A2In2Jn2Jn2-In2=12A1+Jn2A2Jn2A1Jn2-Jn2A2Jn2A1-A2Jn2Jn2A1Jn2+A2.

Since Jn2(A1+Jn2A2Jn2)Jn2=Jn2A1Jn2+A2 and Jn2(A1Jn2-Jn2A2)Jn2=Jn2A1-A2Jn2, then by (Cantoni and Butler 1976, Lemma 2), Fn-1(Q) is centrosymmetric which proves (1). For 1in2,

j=1n(Fn-1(Q))ij=12j=1n(αi,j+βn2+1-i,n2+1-j+αi,n2+1-j-βn2+1-i,j)=j=1nαij.

The above equality means that for 1in2, the i-th row sum of Fn-1(Q) and A1 coincide. This implies that if Fn-1(Q) is a Markov (rate) matrix, then A1 is a Markov (rate) matrix as well. Additionally, note that

(A1+Jn2A2Jn2)ij=αi,j+βn2+1-i,n2+1-jand(A1Jn2-Jn2A2)ij=αi,n2+1-j-βn2+1-i,j.

Hence, (2) and (3) will follow immediately.

Logarithms of centrosymmetric matrices

For the special structure encoded by the centrosymmetric matrices, one may ask whether they have logarithms which are also centrosymmetric. In this section, we provide some answers to this question.

Theorem 3

Let AMn(R) be a CS matrix. Then A has a CS logarithm if and only if both the upper block matrix A1 and the lower block matrix A2 in Lemma 5 admit a logarithm.

Proof

Suppose that A has a centrosymmetric logarithm Q. By Lemma 5, Fn(A)=diag(A1,A2) and Fn(Q)=diag(Q1,Q2). Then exp(Q)=A implies that exp(Q1)=A1 and exp(Q2)=A2. Hence, A1 and A2 admit a logarithm. Conversely, suppose that A1 and A2 admit a logarithm Q1 and Q2, respectively. Then the matrix diag(Q1,Q2) is a logarithm of the matrix diag(A1,A2). By Lemma 6, the matrix Fn-1(diag(Q1,Q2)) is a centrosymmetric logarithm of A.

Proposition 9

Let AMn(R) be a CS matrix. If A is invertible, then it has infinitely many CS logarithms.

Proof

The assumptions imply that the matrices A1 and A2 in Lemma 5 are invertible. By (Higham 2008, Theorem 1.28), each A1 and A2 has infinitely many logarithms. Hence, Theorem 3 implies that A has infinitely many centrosymmetric logarithms.

Proposition 10

Let AMn(R) be a CS matrix such that Log(A) is well-defined. Then Log(A) is again centrosymmetric.

Proof

Let us suppose that Log(A) is not centrosymmetric matrix. Define the matrix Q=Jn(Log(A))Jn. Then QLog(A) since Log(A) is not centrosymmetric. It is also clear that exp(Q)=A. Moreover, since Jn2=In, the matrices Log(A) and Q have the same eigenvalues. Therefore, Q is also a principal logarithm of A, a contradiction to the uniqueness of principal logarithm. Hence, Log(A) must be centrosymmetric.

The following theorem characterizes the logarithms of any invertible CS Markov matrices.

Theorem 4

Let AMn(R) be an invertible CS Markov matrix. Let A1=N1D1N1-1 where D1=diag(R1,R2,,Rl) is a Jordan form of the upper block matrix in Lemma 5. Similarly, let A2=N2D2N2-1 where D2=diag(T1,T2,,Tl) is a Jordan form of the lower block matrix in Lemma 5. Then A has a countable infinitely many logarithms given by

Q:=SnNDN-1Sn-1,

where

N:=diag(N1,N2)andD:=diag(D1,D2),

and Di denotes a logarithm of Di. In particular, these logarithms of A are primary functions of A.

Proof

The theorem follows immediately from (Higham 2008, Theorem 1.28).

For the definition of primary function of a matrix, we refer the reader to Higham (2008). The above theorem says that the logarithms of a nonsingular centrosymmetric matrix contains a countable infinitely many primary logarithms and they are centrosymmetric matrices as well.

Finally, we will present a necessary condition for embeddability of CS Markov matrices in higher dimensions.

Lemma 7

Let n2. Suppose that A=(aij) is an embeddable CS Markov matrix of size n×n with a CS logarithm. Then for n even,

j=1n2(ajj+aj,n-j+1)>1,

while for n odd,

j=1n2(ajj+aj,n-j+1)+an2+1,n2+1>1.

Proof

Since A is an embeddable matrix with CS logarithm, we write A=exp(Q) for some CS rate matrix Q, and then

Fn(A)=Fn(exp(Q))=exp(Fn(Q)).

By Lemma 5, for the centrosymmetric matrices AQ, we have Fn(A)=diag(A1,A2) and Fn(Q)=diag(Q1,Q2) where A1 is a Markov matrix and Q1 is a rate matrix of size n2×n2. Therefore, A1=exp(Q1). If λ1,,λn2 are the eigenvalues, perhaps not distinct, of Q1, then the eigenvalues of A1 are eλ1,,eλn2. Since one of λi’s is zero, then the trace of A1 which is the sum of its eigenvalues is equal to

tr(A1)=j=1n2eλj>1.

We now need to show that trace of A1 has the form written in the lemma. Suppose that n is even. By the proof of Lemma 5, then

tr(A1)=j=1n2(B1+B2Jn2)jj=j=1n2(ajj+aj,n2+j)=j=1n2(ajj+aj,n-j+1).

The proof for odd n can be obtained similarly.

Let XnVnMarkov be the subset containing all centrosymmetric-embeddable Markov matrices. We want to obtain an upper bound of the volume of Xn using Lemma 7. Let YnVnMarkov be the subset containing all centrosymmetric Markov matrices such that after applying the generalized Fourier transformation, the trace of the upper block matrix is greater than 1. The previous lemma implies that XnYn and hence, v(Xn)v(Yn). Moreover, the upper bound v(Yn) is easy to compute as Yn is a polytope and for some values of n, these volumes are presented in Table 8. We see from Table 8, there are at most 50% of matrices in V4Markov that are centrosymmetically-embeddable and hence this upper bound v(Y4) is not good. For n=5, approximately, there are at most 62% in V4Markov that are centrosymmetrically-embeddable but for n=6, this upper bound gives a better proportion, which is approximately 0.1%.

Table 8.

The exact volume v(Yn),n{4,5,6} computed using Polymake

Dimension of VnMarkov v(Yn) v(VnMarkov)
n=4 6 1721.39×10-2 1362.78×10-2
n=5 10 65348384001.35×10-4 146082.17×10-4
n=6 15 4336538371840006.22×10-10 117280005.79×10-7

Embeddability of 6×6 centrosymmetric matrices

Throughout this section we shall consider A to be a 6×6 centrosymmetric Markov matrix with distinct eigenvalues. In particular, the matrices considered in this section are diagonalizable and are a dense subset of all 6×6 centrosymetric Markov matrices. Note that this notation differs from the notation for Markov matrices used in previous sections in order to make it consistent with the notation used in the results presented for generic centrosymmetric matrices.

In the previous section, we showed that F(A) is a block-diagonal real matrix composed of two 3×3 blocks denoted by A1 and A2. Since both A1 and A2 have real entries, each of these matrices has at most one conjugate pair of eigenvalues. Adapting the notation introduced in Theorem 4 to diagonalizable matrices we have N1,N2GL3(C) such that A1=N1diag(1,λ1,λ2)N1-1 and A2=N2diag(μ,γ1,γ2)N2-1 with μR>0 and λi,γiC\R0. Moreover, we can assume that Im(λ1)>0 without loss of generality (this can be achieved by permuting the second and third columns of N1 if necessary). For ease of reading, we will define as P:=S6diag(N1,N2), where S6 is the matrix used to obtain the Fourier transform F(A) and was introduced in Sect. (5.3).

Next we give a criterion for the embeddability of A for each of the following cases: graphic file with name 285_2023_1895_Figa_HTML.jpg

Proposition 11

If a 6×6 cetrosymmetric Markov matrix A does not belong to any of the cases in Table 6.1, then it is not embeddable.

Proof

If A satifies the hypothesis of the proposition then either it has a null eigenvalue or it has a simple negative eigenvalue. In the former case A is a singular matrix and hence it has no logarithm. If A had a simple negative eigenvalue, then all its logarithms would have a non-real eigenvalue whose complementary pair is not an eigenvalue of A (otherwise M would have a repeated eigenvalues). Therefore, A has no real logarithm.

Remark 7

All the results in this section can be adapted to 5×5 centrosymmetric Markov matrices by not considering the eigenvalue μ and modifying the forthcoming definitions of the matrices Log-1(A) and V accordingly (i.e. removing the fourth row and column in the corresponding diagonal matrix). In addition, these results still hold if the eigenvalue 1 of the Markov matrix has multiplicity 2.

Case 1

The results for this case are not restricted to centrosymmetric matrices but can be applied to decide the embeddability of any suitable Markov matrix.

Proposition 12

If all the eigenvalues of a Markov matrix A are distinct and positive, then A is embeddable if and only if Log(A) is a rate matrix.

Proof

If A has distinct real eigenvalues then it has only one real logarithm, which is Log(A) (see (Culver 1966)).

Case 2

In this case A has exactly one conjugate pair of complex eigenvalues and we obtain the following criterion by adapting Corollary 5.6 in Casanellas et al. (2023) to our framework:

Proposition 13

Given the matrix V:=Pdiag(0,0,0,0,2πi,-2πi)P-1 define:

graphic file with name 285_2023_1895_Equ98_HTML.gif

and set N:={(i,j):ij,Vi,j=0andLog(A)i,j<0}. Then,

  1. A is embeddable if and only if N= and LU.

  2. the set of Markov generators for A is {Q=Log(A)+kV:kZsuch thatLkU}.

Proof

The proof of this theorem is analogous to the proof of Theorem 5.5 in Casanellas et al. (2020a) but considering the matrix V as defined here. According to Proposition 1, any Markov generator of A is of the form

Logk(A)=Pdiag(0,log(λ1),log(λ2),log(μ),logk(γ1),logk(γ1)¯)P-1=Pdiag(0,log(λ1),log(λ2),log(μ),logk(γ1)+2πki,logk(γ1)¯-2πki)P-1.

Such a logarithm can be rewritten as Log(A)+kV. Using this, we will prove that Logk(A)=Log(A)+kV is a rate matrix if and only if N= and LkU.

Suppose that there exists kZ such that Logk(A) is a rate matrix. Hence, Log(A)i,j+kVi,j0 for all ij. For ij, we have:

  1. Log(A)i,j0 for all ij such that Vi,j=0. This means that N=.

  2. -Log(A)i,jVi,jk for all ij such that Vi,j>0. This means that Lk.

  3. -Log(A)i,jVi,jk for all ij such that Vi,j<0. This means that kU.

Conversely, suppose that N= and and that there is kZ such that LkU. We want to check that Logk(A) is a rate matrix. According to Proposition 1, each row of Logk(A) sums to 0. Moreover, for ij, we have:

  1. if Vi,j=0, then Logk(A)i,j=Log(A)i,j. Since N=, Logk(A)i,j=Log(A)i,j0.

  2. if Vi,j>0, then Logk(A)i,j=Log(A)i,j+kVi,jLog(A)i,j+LVi,jLog(A)i,j+(-Log(A)i,jVi,j)Vi,j=0.

  3. if Vi,j<0, then -Logk(A)i,j=-Log(A)i,j-kVi,j-Log(A)i,j-UVi,j-Log(A)i,j-(-Log(A)i,jVi,j)Vi,j=0.

The proof is now complete.

Case 3

As in Case 2, A has exactly one conjugate pair of eigenvalues and hence its embeddability (and all its generators) can be determined by using Proposition 13 but defining the matrix V as V=Pdiag(0,0,0,0,2πi,-2πi)P-1. However in Case 3 the conjugate pair of eigenvalues lie in A1 which is a Markov matrix. This allows us to use the results regarding the embeddability of 3×3 Markov matrices to obtain an alternative criterion to test the embeddability of A. To this end we define

Log-1(A):=Pdiag(0,z,z¯,log(μ),log(γ1)log(γ2))P-1 6.2

where z:=log-1(λ1).

Proposition 14

The matrix A is embeddable if and only if Log(A) or Log-1(A) are rate matrices.

Proof

Note that exp(Log(A))=exp(Log-1(A))=A so one of the implications is immediate to prove. To prove the other implication, we assume that A is embeddable and let Q be a Markov generator for it. Proposition 1 yields that

Q=Pdiag(0,logk1(λ1),logk2(λ2),logk3(μ),logk4(γ1),logk5(γ2))P-1,

for some integers k1,,k5Z. Therefore, F(Q)=Q100Q2 where Q1 and Q2 are real logarithms of A1 and A2 respectively.

Since A2 is a real matrix with distinct positive eigenvalues, its only real logarithm is its principal logarithm. This implies that k3=k4=k5=0 (so that Q2=Log(A2)).

Now, recall that A1 is a Markov matrix (see Lemma 5). Using Proposition 1 again, we obtain that Q1 is a rate matrix, thus A1 is embeddable. To conclude the proof it is enough to recall Theorem 4 in James (1973), which yields that A1 is embeddable if and only if Log(A1) or P1diag(0,z,z¯)P1-1 is a rate matrix.

Case 4

In this case, the solution to the embedding problem can be obtained as a byproduct of the results for the previous cases:

Proposition 15

Let Log0,0(A) denote the principal logarithm of A and Log-1,0(A) denote the matrix in (6.2). Given the matrix V:=Pdiag(0,0,0,0,2πi,-2πi)P-1 and k{0,-1} define:

graphic file with name 285_2023_1895_Equ99_HTML.gif

and set Nk:={(i,j):ij,Vi,j=0andLogk,0(A)i,j<0}. Then,

  1. A is embeddable if and only if Nk= and LkUk for k=0 or k=-1.

  2. If A is embeddable, then at least one of its Markov generator can be written as
    Logk,k2(A):=Pdiag(0,logk(λ1),logk(λ1)¯,log(μ),logk2(γ1),logk2(γ1)¯)P-1
    with k{0,-1} and k2Z such that Lkk2Uk.

Proof

The matrix A is embeddable if and only if it admits a Markov generator. According to Proposition 1, if such a generator Q exists then it can be written as Logk1,k2(A) for some k1,k2Z. Therefore, Lemma 5 implies that F(A)=A100A2 for some matrices A1 and A2. Moreover, F(Q)=Q100Q2 where Q1 and Q2 are real logarithms of A1 and A2 respectively.

As shown in the proof of Proposition 14, A1 is actually a Markov matrix and Q1 is a Markov generator for it (see also Lemma 5). Moreover, by Theorem 4 in James (1973), A1 is embeddable if and only if Log(A1) or Log-1(A1) are rate matrices. This implies that Logk1,k2(A) is a rate matrix if and only if Log0,k2(A) or Log-1,k2 are rate matrices. To conclude the proof we proceed as in the proof of Proposition 13. Indeed, note that for k{0,-1}, Logk,k2(A)=Logk,0(A)+k2V. Using this, it is immediate to check that Logk,k2(A) is a rate matrix if and only if Nk= and Lkk2Uk.

Discussion

The central symmetry is motivated by the complementarity between both strands of the DNA. When a nucleotide substitution occurs in one strand, there is also a substitution between the corresponding complementary nucleotides on the other strand. Therefore, working with centrosymmetric Markov matrices is the most general approach when considering both DNA strands.

In this paper, we have discussed the embedding problem for centrosymmetric Markov matrices. In Theorem 2, we have obtained a characterization of the embeddabilty of 4×4 centrosymmetric Markov matrices which are exactly the strand symmetric Markov matrices. In particular, we have also shown that if a 4×4 CS Markov matrix is embeddable, then any of its Markov generators is also a CS matrix. Furthermore, In Sect. 6, we have discussed the embeddability criteria for larger centrosymmetric matrices.

As a consequence of the characterization of Theorem 2, we have been able to compute and compare the volume of the embeddable 4×4 CS Markov matrices within some subspaces of 4×4 CS Markov matrices. These volume comparisons can be seen in Table 2 and Table 7. For larger matrices, using the results in Sect. 6, we have estimated the proportion of embeddable matrices within the set of all 6×6 centrosymmetric Markov matrices and within the subsets of DLC and DD matrices. This is summarized in Table 9 below. The computations were repeated several times obtaining results with small differences in the values but the same order of magnitude and starting digits.

Table 9.

Relative volume of embeddable matrices within relevant subsets of 6×6 centrosymmetric Markov matrices. The results were obtained using the hit-and-miss Monte Carlo integration with 107 sample points

Set Sample points Embeddable sample points Rel. vol. of embeddable matrices
V6Markov 108 1370 0.0000137
VDLC 1034607 1362 0.0013164
VDD 3048 84 0.0275590

As we have seen in Sect. 3 and 6, we have only considered in detail the embeddability of CS Markov matrices of size n=4 and n=6. We expect that the proportion of the embeddable CS Markov matrices within the subset of Markov matrices in larger dimension tends to zero as n grows larger as indicated by Tables 2, 78, and 9.

These results together with the results obtained for the strand symmetric model (see Table 7) indicate that restricting to homogeneous Markov processes in continuous-time is a very strong restriction because non-embeddable matrices are discarded and their proportion is much larger than that of embeddable matrices. For instance, in the 2×2 case exactly 50% of the matrices are discarded (Ardiyansyah et al. 2021, Table 5), while in the case of 4×4 matrices up to 98.26545% of the matrices are discarded (see Table 7) and in the case of 6×6 matrices the amount of discarded matrices is about 99.99863% as indicated in Table 9. However, when restricting to subsets of Markov matrices which are mathematically more meaningful in biological terms, such as DD or DLC matrices, the proportion of embeddable matrices is much higher so that we are discarding less matrices (e.g. for DD we discard 68.41679% of 4×4 matrices and 97.2441% of 6×6 matrices). This is not to say that it makes no sense to use continuous-time models but to highlight that one should take the above restrictions into consideration when working with these models. Conversely, when working with the whole set of Markov matrices one has to be aware that they might end up considering lots of non-meaningful matrices.

Acknowledgements

Dimitra Kosta was partially supported by a Royal Society Dorothy Hodgkin Research Fellowship DHF\R1\201246. Jordi Roca-Lacostena was partially funded by Secretaria d’Universitats i Recerca de la Generalitat de Catalunya (AGAUR 2018FI_B_0094). Muhammad Ardiyansyah is partially supported by the Academy of Finland Grant No. 323416.

Funding Information

Open Access funding provided by Aalto University.

Declarations

Conflict of interest

The authors declare no conflict of interest.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Muhammad Ardiyansyah, Email: muhammad.ardiyansyah@aalto.fi.

Dimitra Kosta, Email: D.Kosta@ed.ac.uk.

Jordi Roca-Lacostena, Email: jordi.roca.lacostena@upc.edu.

References

  1. Aitken AC (2017) Determinants and matrices. Read Books Ltd
  2. Ardiyansyah M, Kosta D, Kubjas K. The model-specific Markov embedding problem for symmetric group-based models. J Math Biol. 2021;83(3):1–26. doi: 10.1007/s00285-021-01656-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Baake M, Sumner J. Notes on markov embedding. Linear Algebra Appl. 2020;594:262–299. doi: 10.1016/j.laa.2020.02.016. [DOI] [Google Scholar]
  4. Bain J, Switzer C, Chamberlin R, Benner SA. Ribosome-mediated incorporation of a non-standard amino acid into a peptide through expansion of the genetic code. Nature. 1992;356(6369):537–539. doi: 10.1038/356537a0. [DOI] [PubMed] [Google Scholar]
  5. Benner SA, Sismour AM. Synthetic Biol. Nat Rev Genet. 2005;6(7):533–543. doi: 10.1038/nrg1637. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Cantoni A, Butler P. Eigenvalues and eigenvectors of symmetric centrosymmetric matrices. Linear Algebra Appl. 1976;13(3):275–288. doi: 10.1016/0024-3795(76)90101-4. [DOI] [Google Scholar]
  7. Carette P. Characterizations of embeddable 3×3 stochastic matrices with a negative eigenvalue. New York J Math. 1995;1:120–129. [Google Scholar]
  8. Casanellas M, Kedzierska AM. Generating Markov evolutionary matrices for a given branch length. Linear Algebra Appl. 2013;438(5):2484–2499. doi: 10.1016/j.laa.2012.11.009. [DOI] [Google Scholar]
  9. Casanellas M, Sullivant S (2005) The strand symmetric model. In: Pachter L, Sturmfels B (eds) Algebraic statistics for computational biology. Cambridge University Press, New York
  10. Casanellas M, Fernández-Sánchez J, Roca-Lacostena J. Embeddability and rate identifiability of Kimura 2-parameter matrices. J Math Biol. 2020;80(4):995–1019. doi: 10.1007/s00285-019-01446-0. [DOI] [PubMed] [Google Scholar]
  11. Casanellas M, Fernández-Sánchez J, Roca-Lacostena J (2020b) An open set of 4× 4 embeddable matrices whose principal logarithm is not a markov generator. Linear and Multilinear Algebra pp 1–12
  12. Casanellas M, Fernández-Sánchez J, Roca-Lacostena J. The embedding problem for Markov matrices. Publicacions Matemàtiques. 2023;67:411–445. doi: 10.5565/PUBLMAT6712308. [DOI] [Google Scholar]
  13. Chang JT. Full reconstruction of Markov models on evolutionary trees: identifiability and consistency. Math Biosci. 1996;137(1):51–73. doi: 10.1016/S0025-5564(96)00075-2. [DOI] [PubMed] [Google Scholar]
  14. Chen Y, Chen J. On the imbedding problem for three-state time homogeneous Markov chains with coinciding negative eigenvalues. J Theor Probab. 2011;24:928–938. doi: 10.1007/s10959-010-0316-5. [DOI] [Google Scholar]
  15. Culver WJ. On the existence and uniqueness of the real logarithm of a matrix. Proceed Am Math Soc. 1966;17:1146–1151. doi: 10.1090/S0002-9939-1966-0202740-6. [DOI] [Google Scholar]
  16. Cuthbert JR. On uniqueness of the logarithm for Markov semi-groups. J Lond Math Soc. 1972;2(4):623–630. doi: 10.1112/jlms/s2-4.4.623. [DOI] [Google Scholar]
  17. Davies E, et al. Embeddable Markov matrices. Electron J Probab. 2010;15:1474–1486. doi: 10.1214/EJP.v15-733. [DOI] [Google Scholar]
  18. Elfving G (1937) Zur theorie der Markoffschen ketten. Acta Societatis Scientiarum FennicæNova Series A 2(8):17 pages
  19. Fuglede B. On the imbedding problem for stochastic and doubly stochastic matrices. Probab Theory Relat Fields. 1988;80:241–260. doi: 10.1007/BF00356104. [DOI] [Google Scholar]
  20. Gawrilow E, Joswig M (2000) Polymake: a framework for analyzing convex polytopes. In: Polytopes-combinatorics and computation, Springer, pp 43–73
  21. Goodman GS. An intrinsic time for non-stationary finite Markov chains. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete. 1970;16:165–180. doi: 10.1007/BF00534594. [DOI] [Google Scholar]
  22. Hammersley J (2013) Monte carlo methods. Springer Science & Business Media
  23. Higham NJ (2008) Functions of matrices: Theory and computation, vol 104. SIAM
  24. Hoshika S, Leal NA, Kim MJ, Kim MS, Karalkar NB, Kim HJ, Bates AM, Watkins NE, SantaLucia HA, Meyer AJ, et al. Hachimoji DNA and RNA: a genetic system with eight building blocks. Science. 2019;363(6429):884–887. doi: 10.1126/science.aat0971. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Inc WR (2022) Mathematica, Version 13.1 https://www.wolfram.com/mathematica
  26. Iosifescu M (2014) Finite Markov processes and their applications. Courier Corporation
  27. James CR. The logarithm function for finite-state Markov semi-groups. J Lond Math Soc. 1973;2(3):524–532. [Google Scholar]
  28. Jia C. A solution to the reversible embedding problem for finite markov chains. Statist Probab Lett. 2016;116:122–130. doi: 10.1016/j.spl.2016.04.020. [DOI] [Google Scholar]
  29. Johansen S (1974) Some results on the imbedding problem for finite Markov chains. J London Math Soc s2-8(2):345–351
  30. Kimura M (1957) Some problems of stochastic processes in genetics. Ann Math Statistics pp 882–901
  31. Kingman JFC. The imbedding problem for finite Markov chains. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete. 1962;1(1):14–24. doi: 10.1007/BF00531768. [DOI] [Google Scholar]
  32. Leal NA, Kim HJ, Hoshika S, Kim MJ, Carrigan MA, Benner SA. Transcription, reverse transcription, and analysis of RNA containing artificial genetic components. ACS Synth Biol. 2015;4(4):407–413. doi: 10.1021/sb500268n. [DOI] [PubMed] [Google Scholar]
  33. Malyshev DA, Dhami K, Quach HT, Lavergne T, Ordoukhanian P, Torkamani A, Romesberg FE (2012) Efficient and sequence-independent replication of DNA containing a third base pair establishes a functional six-letter genetic alphabet. Proceedings of the National Academy of Sciences 109(30):12,005–12,010 [DOI] [PMC free article] [PubMed]
  34. Pachter L, Sturmfels B. Algebraic statistics for computational biology, UK: Cambridge University Press; 2005. [Google Scholar]
  35. Roca-Lacostena J (2021) The embedding problem for Markov matrices. PhD thesis, Universitat Politècnica de Catalunya
  36. Roca-Lacostena J, Fernández-Sánchez J. Embeddability of Kimura 3ST Markov matrices. J Theor Biol. 2018;445:128–135. doi: 10.1016/j.jtbi.2018.02.005. [DOI] [PubMed] [Google Scholar]
  37. Roca-Lacostena J, Fernández-Sánchez J. Embeddability of Kimura 3st markov matrices. J Theor Biol. 2018;445:128–135. doi: 10.1016/j.jtbi.2018.02.005. [DOI] [PubMed] [Google Scholar]
  38. Runnenberg JT. On Elfving’s problem of imbedding a time-discrete markov chain in a time-continuous one for finitely many states. Proceedings of the KNAW - Series A, Mathematical Sciences. 1962;65:536–541. [Google Scholar]
  39. Schensted IV. Appendix model of subnuclear segregation in the macronucleus of ciliates. Am Nat. 1958;92(864):161–170. doi: 10.1086/282023. [DOI] [Google Scholar]
  40. Sismour AM, Lutz S, Park JH, Lutz MJ, Boyer PL, Hughes SH, Benner SA. Pcr amplification of DNA containing non-standard base pairs by variants of reverse transcriptase from human immunodeficiency virus-1. Nucleic Acids Res. 2004;32(2):728–735. doi: 10.1093/nar/gkh241. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Stein P. A note on the volume of a simplex. Am Math Mon. 1966;73(3):299–301. doi: 10.2307/2315353. [DOI] [Google Scholar]
  42. Weaver JR. Centrosymmetric (cross-symmetric) matrices, their basic properties, eigenvalues, and eigenvectors. Am Math Mon. 1985;92(10):711–717. doi: 10.1080/00029890.1985.11971719. [DOI] [Google Scholar]
  43. Yang Z, Hutter D, Sheng P, Sismour A, Benner S. Artificially expanded genetic information system: a new base pair with an alternative hydrogen bonding pattern. Nucleic Acids Res. 2006;34(21):6095–101. doi: 10.1093/nar/gkl633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Yang Z, Sismour AM, Sheng P, Puskar NL, Benner SA. Enzymatic incorporation of a third nucleobase pair. Nucleic Acids Res. 2007;35(13):4238–4249. doi: 10.1093/nar/gkm395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Yang Z, Chen F, Alvarado JB, Benner SA (2011) Amplification, mutation, and sequencing of a six-letter synthetic genetic system. J Am Chem Soc 133(38):15,105–15,112 [DOI] [PMC free article] [PubMed]
  46. Yap VB, Pachter L. Identification of evolutionary hotspots in the rodent genomes. Genome Res. 2004;14(4):574–579. doi: 10.1101/gr.1967904. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Journal of Mathematical Biology are provided here courtesy of Springer

RESOURCES