Abstract
In this paper, we discuss the embedding problem for centrosymmetric matrices, which are higher order generalizations of the matrices occurring in strand symmetric models. These models capture the substitution symmetries arising from the double helix structure of the DNA. Deciding whether a transition matrix is embeddable or not enables us to know if the observed substitution probabilities are consistent with a homogeneous continuous time substitution model, such as the Kimura models, the Jukes-Cantor model or the general time-reversible model. On the other hand, the generalization to higher order matrices is motivated by the setting of synthetic biology, which works with different sizes of genetic alphabets.
Keywords: Evolutionary model, Embedding problem, Markov matrix, Centrosymmetric matrix
Introduction
Phylogenetics is the study of evolutionary relationships among biological entities, also known as taxa, that aims to infer the evolutionary history among them. In order to model evolution, we consider a directed acyclic graph, called a phylogenetic tree, depicting the evolutionary relationships amongst a selected set of taxa. Phylogenetic trees consist of vertices and edges. Vertices represent taxa, while edges between vertices represent the evolutionary processes between the taxa.
In order to describe the real evolutionary process along an edge of a phylogenetic tree, one often assumes that the evolutionary data occurred following a Markov process. A Markov process is a random process in which the future is independent of the past, given the present. Under this Markov process, transitions between n states given by conditional probabilities are presented in a Markov matrix M, that is a square matrix whose entries are nonnegative and rows sum to one. A well-known problem in probability theory is the so-called embedding problem which was initially posed by Elfving (Elfving 1937). The embedding problem asks whether given a Markov matrix M, one can find a real square matrix Q with rows summing to zero and non-negative off-diagonal entries, such that . The matrix Q is called a Markov generator.
In the complex setting, the embedding problem is completely solved by Higham (2008); a complex matrix A is embeddable if and only if A is invertible. However, as our motivation arises from molecular models of evolution we are interested in the embedding problem over the real numbers, so from now on we will denote by M a real Markov matrix. It was shown by Kingman (1962) that if an real Markov matrix M is embeddable, then the matrix M has . Moreover, in the same work by Kingman it was shown that is a necessary and sufficient condition for a Markov matrix M to be embeddable. For Markov matrices a complete solution of the embedding problem is provided in a series of papers (James 1973; Johansen 1974; Carette 1995; Chen and Chen 2011), where the characterisation of embeddable matrices depends on the Jordan decomposition of the Markov matrix. For Markov matrices the embedding problem is completely settled in a series of papers (Casanellas et al. 2020a, 2023; Roca-Lacostena and Fernández-Sánchez 2018a), where similarly to the case the full characterisation of embeddable matrices is distinguished into cases depending on the Jordan form of the Markov matrices.
For the general case of Markov matrices, there are several results; some presenting necessary conditions (Elfving 1937; Kingman 1962; Runnenberg 1962), while others sufficient conditions (James 1973; Fuglede 1988; Goodman 1970; Davies et al. 2010) for embeddability of Markov matrices. Moreover, the embedding problem has been solved for special matrices with a biological interest such as equal-input and circulant matrices (Baake and Sumner 2020), group-based models (Ardiyansyah et al. 2021) and time-reversible models (Jia 2016). Despite the fact that there is no theoretical explicit solution for the embeddability of general Markov matrices, there are results (Casanellas et al. 2023) that enable us to decide whether a Markov matrix with distinct eigenvalues is embeddable or not. This is achieved by providing an algorithm that outputs all Markov generators of such a Markov matrix (Casanellas et al. 2023; Roca-Lacostena 2021).
In this paper, we focus on the embedding problem for matrices that are symmetric about their center and are called centrosymmetric matrices (see Definition 2). We also study a variation of the famous embedding problem called model embeddability, where apart from the requirement that the Markov matrix is the matrix exponential of a rate matrix, we additionally ask that the rate matrix follows the model structure. For instance, for centrosymmetric matrices, model embeddability means that the rate matrix is also centrosymmetric.
The motivation for studying centrosymmetric matrices comes from evolutionary biology, as the most general nucleotide substitution model when considering both DNA strands admits any centrosymmetric Markov matrix as a transition matrix, where n is the even number of nucleotides. For instance, by considering the four natural nucleotides A-T, C-G we arrive at the strand symmetric Markov model, a well-known phylogenetic model whose substitution probabilities reflect the symmetry arising from the complementarity between the two strands that the DNA is composed of (see (Casanellas and Sullivant 2005)). In particular, a strand symmetric model for DNA must have the following equalities of probabilities in the root distribution:
| 1.1 |
and the following equalities of probabilities in the transition matrices
Therefore, the corresponding transition matrices of this model are centrosymmetric matrices, usually called strand symmetric Markov matrices in this context. In this article, we will use the terminology centrosymmetric Markov matrix and strand symmetric Markov matrix interchangeably. In the strand symmetric model there are less restrictions on the way genes mutate from ancestor to child compared to other widely known molecular models of evolution. In fact, special cases of the strand symmetric model are the group-based phylogenetic models such as the Jukes-Cantor (JC) model, the Kimura 2-parameter (K2P) and Kimura 3-parameter (K3P) models. The algebraic structure of strand symmetric models was initially studied in (Casanellas and Sullivant 2005), where it was argued that strand symmetric models capture more biologically meaningful features of real DNA sequences than the commonly used group-based models, as for instance, in any group-based model, the stationary distribution of bases for a single species is always the uniform distribution, while computational evidence in (Yap and Pachter 2004) suggests that the stationary distribution of bases for a single species is rarely uniform, but must always satisfy the symmetries (1.1) arising from nucleotide complementarity, as assumed by the strand symmetric model.
In this article, we also explore higher order centrosymmetric matrices for which , which is justified by the use of synthetic nucleotides. One of main goals of synthetic biology is to expand the genetic alphabet to include an unnatural or synthetic base pair. The more letters in a genetic system could possibly lead to an increased potential for retrievable information storage and bar-coding and combinatorial tagging (Benner and Sismour 2005). Naturally the four-letter genetic alphabet consists of just two pairs, A-T and G-C. In 2012, a genetic system comprising of three base pairs was introduced in (Malyshev et al. 2012). In addition to the natural base pairs, the third, unnatural or synthetic base pair 5SICS-MMO2 was proven to be functionally equivalent to a natural base pair. Moreover, when it is combined with the natural base pairs, 5SICS-MMO2 provides a fully functional six-letter genetic alphabet. Namely, six-letter genetic alphabets can be copied (Yang et al. 2007), polymerase chain reaction (PCR)-amplified and sequenced (Sismour et al. 2004; Yang et al. 2011), transcribed to six-letter RNA and back to six-letter DNA (Leal et al. 2015), and used to encode proteins with added amino acids (Bain et al. 1992). This biological importance and relevance of the above six-letter genetic alphabets motivates us to particularly study the Markov matrices describing the probabilities of changing base pairs in the six-letter genetic system in Sect. 6. When considering both DNA strands, each substitution is observed twice due to the complementarity between both strands, and hence the resulting transition matrix is centrosymmetric.
Moreover there are other synthetic analogs to natural DNA which justify studying centrosymmetric matrices for . For instance, hachimoji DNA is a synthetic DNA that uses four synthetic nucleotides B,Z,P,S in addition to the four natural ones A,C,G,T. With the additional four synthetic ones, hachimoji DNA forms four types of base pairs, two of which are unnatural: P binds with Z and B binds with S. The complementarity between both strands of the DNA implies that the transition matrix is centrosymmetric. Moreover, the research group responsible for the hachimoji DNA system had also studied a synthetic DNA analog system that used twelve different nucleotides, including the four found in DNA (see Yang et al. 2006). Although the biological models which motivate the study of centrosymmetric matrices in this paper require n to be an even number due to the double-helix structure of DNA, in Sect. 5, we include the case of n being odd for completeness.
Apart from embeddability, that is existence of Markov generators, it is also natural to ask about uniqueness of a Markov generator which is called the rate identifiability problem. Identifiability is a property which a model must satisfy in order for precise statistical inference to be possible. A class of phylogenetic models is identifiable if any two models in the class produce different data distributions. In this article, we further develop the results on rate identifiability of the Kimura two parameter model (Casanellas et al. 2020a) to study rate identifiability for strand symmetric models. We also show that there are embeddable strand symmetric Markov matrices with non identifiable rates, namely the Markov generator is not unique. Moreover, we show that strand symmetric Markov matrices are not generically identifiable, that is, there exists a positive measure subset of strand symmetric Markov matrices containing embeddable matrices whose rates are not identifiable.
This paper is organised as follows. In Sect. 2, we introduce the basic definitions and results on embeddability. In Sect. 3, we give a characterisation for a centrosymmetric Markov matrix M with four distinct real nonnegative eigenvalues to be embeddable providing necessary and sufficient conditions in Theorem 2, while we also discuss their rate identifiability property in Proposition 3. Moreover in Sect. 4, using the conditions of our main result Theorem 2, we compute the relative volume of all centrosymmetric Markov matrices relative to the centrosymmetric Markov matrices with positive eigenvalues and , as well as the relative volume of all centrosymmetric Markov matrices relative to the centrosymmetric Markov matrices with four distinct eigenvalues and . We also compare the results on relative volumes obtained using our method with the algorithm suggested in Casanellas et al. (2023) to showcase the advantages of our method. In Sect. 5, we study higher order centrosymmetric matrices and motivate their use in Sect. 6 by exploring the case of synthetic nucleotides where the phylogenetic models admit centrosymmetric mutation matrices. Finally, Sect. 7 discusses implications and possibilities for future work.
Preliminaries
In this section we will introduce the definitions and results that will be required throughout the paper. We will denote by the set of square matrices with entries in the field . The subset of non-singular matrices in will be denoted by .
Definition 1
A Markov (or transition) matrix is a non-negative real square matrix with rows summing to one. A rate matrix is a real square matrix with rows summing to zero and non-negative off-diagonal entries.
In this paper, we are focusing on a subset of Markov matrices called centrosymmetric Markov matrices.
Definition 2
A real matrix is said to be centrosymmetric (CS) if
for every .
Definition 2 reveals that a CS matrix is nothing more than a square matrix which is symmetric about its center. This class of matrices has been previously studied, for instance, in (Aitken 2017, page 124) and Weaver (1985). Examples of CS matrices for and , are the following two matrices respectively:
The class of CS matrices plays an important role in the study of Markov processes since they are indeed transition matrices for some processes in evolutionary biology. For instance, in Kimura (1957), centrosymmetric matrices are used to study the random assortment phenomena of subunits in chromosome division. Furthermore, in Schensted (1958), the same centrosymmetric matrices appear as the transition matrices in the model of subnuclear segregation in the macronucleus of ciliates. Finally, the work (Iosifescu 2014) examines a special case of the random genetic drift phenomenon, which consists of a population of individuals that are able to produce a single type of gamete. In this case, the transition matrices of the associated Markov chain are given by centrosymmetric matrices.
The embedding problem is directly related to the notions of matrix exponential and logarithm which we introduce for completeness below.
Definition 3
We define the exponential of a matrix A, using the Taylor power series of the function , as
where and denotes the identity matrix. If is an eigendecomposition of A, then . Given a matrix , a matrix is said to be a logarithm of A if . If v is an eigenvector corresponding to the eigenvalue of A, then v is an eigenvector corresponding to the eigenvalue of .
A Markov matrix M is called embeddable if it can be written as the exponential of a rate matrix Q, namely . Then any rate matrix Q satisfying the equation is called a Markov generator of M.
Remark 1
Embeddable Markov matrices occur when we assume a continuous time Markov chain, in which case the Markov matrices have the form
where represents time and Q is a rate matrix. However, in the rest of the paper, we assume that t is incorporated in the rate matrix Q.
The existence of multiple logarithms is a direct consequence of the distinct branches of the logarithmic function in the complex field.
Definition 4
Given and , the k-th branch of the logarithm of z is , where is the logarithmic function on the real field and denotes the principal argument of z. The logarithmic function arising from the branch is called the principal logarithm of z and is denoted as .
It is known that if A is a matrix with no negative eigenvalues, then there is a unique logarithm of A all of whose eigenvalues are given by the principal logarithm of the eigenvalues of A (Higham 2008, Theorem 1.31). We refer to this unique logarithm as the principal logarithm of A, denoted by Log(A).
By definition, the Markov generators of a Markov matrix M are those logarithms of M that are rate matrices. In particular they are real logarithms of M. The following result enumerates all the real logarithms with rows summing to zero of any given Markov matrix with positive determinant and distinct eigenvalues. Therefore, all Markov generators of such a matrix are necessarily of this form.
Proposition 1
(Casanellas et al. 2023, Proposition 4.3). Let be an Markov matrix with and distinct eigenvalues for and for , all of them pairwise distinct. Then, a matrix Q is a real logarithm of M with rows summing to zero if and only if for some .
Remark 2
In particular, the principal logarithm of M can be computed as
In this paper, we focus on the embedding problem for the class of centrosymmetric matrices. In Sect. 3, we will first study the embeddability of centrosymmetric Markov matrices, which include the K3P, K2P and JC Markov matrices. In Sect. 5 and Sect. 6, we will further study the embeddability of higher order centrosymmetric Markov matrices.
Embeddability of centrosymmetric matrices
In this section, we begin our study by analyzing the embeddability of centrosymmetric matrices also known as strand symmetric matrices. We will provide necessary and sufficient conditions for centrosymmetric matrices to be embeddable. Moreover, we will discuss their rate identifiability problem as well.
The transition matrices of centrosymmetric matrices are assumed to have the form
where
Recall that the K3P matrices are assumed to have the form
In the case of the K2P matrices, we additionally have , while in the case of JC matrices, . It can be easily seen that K3P, K2P, and JC Markov (rate) matrices are centrosymmetric.
Let us define the following matrix
| 3.1 |
compare (Casanellas and Kedzierska 2013, Section 6). For a CS Markov matrix M, we define . By direct computation, it can be checked that F(M) is a block diagonal matrix
| 3.2 |
where
| 3.3 |
Define two matrices, and , which are the upper and lower block matrices in (3.2), respectively.
Similarly, the rate matrices in strand symmetric models are assumed to have the centrosymmetric form
where
So, for a CS rate matrix Q, we can also define . By direct computation, it can be checked that
| 3.4 |
where
Define two matrices, and , which are the upper and lower block matrices in (3.4), respectively.
The following results provide necessary conditions for a CS Markov matrix to be embeddable.
Lemma 1
Let be a CS Markov matrix and for some CS rate matrix Q. Then
and
Proof
We have that
Then
Thus, is an embeddable Markov matrix. Using the embeddability criteria of Markov matrices in Kingman (1962), we have that , which is the desired inequality. Additionally, since , as desired.
Lemma 2
Let be a CS Markov matrix and for some CS rate matrix . If , then
and
Proof
By direct computations and the proof of Lemma 1,
We then have the following system of equations:
| 3.5 |
Summing the two equations, we get
Note that by Lemma 1, Therefore,
| 3.6 |
Using Equation (3.5) and (3.6), we obtain
The proof is now complete.
Proposition 2
Given two matrices , consider the block-diagonal matrix . Then the following statements hold:
-
i)
is a CS matrix.
-
ii)is a Markov matrix if and only if A is a Markov matrix and
-
iii)is a rate matrix if and only if A is a rate matrix and
Proof
To prove i), by direct computation we obtain that
Then ii) follows from the above expression of and the fact that rows of Markov matrices add to 1 and the entries are non-negative, while iii) similarly follows from the fact that the rows of rate matrices add to zero and the off-diagonal entries are non-negative.
For any CS Markov matrix , let us recall that by (3.2), M is block-diagonalizable via the matrix S. In the rest of this section, we will study both the upper and the lower block matrices of F(M) more closely. Studying the upper and lower blocks allows us to establish the main result of the embeddability criteria for CS Markov matrices. This block-diagonalization reduces our analysis to studying the logarithms of both the upper and the lower block matrices which have size . This result will be presented in Theorem 2.
Upper block
As we have seen in (3.2), the upper block of F(M) is given by the matrix , which is a Markov matrix. If , then
Hence, by Proposition 1, any logarithm of can be written as
for some integers and . Let , , and . Then
| 3.7 |
Lemma 3
If , then is a real matrix if and only if and . In this case, the only real logarithm of is the principal logarithm
Proof
For fixed and , the eigenvalues of are and Then is a real matrix if and only if or . Since , . Thus, is a real matrix if and only if . Finally, if and only if and if and only if and .
Lower block
The lower block of F(M) is given by the matrix . Unlike , the matrix is generally not a Markov matrix. The discriminant of the characteristic polynomial of is given by
| 3.8 |
with defined as in (3.3). If , then has two distinct real eigenvalues and if , then has a pair of conjugated complex eigenvalues. Moreover, if , then has either Jordan block or a repeated real eigenvalue. We will assume that so that diagonalizes into two distinct eigenvalues.
Let . Then
Let us now define
where and are integers. Therefore, any logarithm of can be written as
| 3.9 |
where
Lemma 4
If , then is a real matrix if and only if and
If , then is a real matrix if and only if .
Proof
If , then and . Moreover, Since and are the eigenvalues of , this implies that . In particular, is a real matrix if and only if both and are real.
Let us assume and take . Fixing , the eigenvalues of are and , which are both complex numbers. Thus, is real if and only if . Hence, Conversely, implies that and . Thus, all entries of are real.
Logarithms of CS Markov matrices
Let M be a CS Markov matrix. Using the values defined in (3.3) and (3.8), we can now label up its four eigenvalues as following,
| 3.10 |
We note that the subset of CS Markov matrix with repeated eigenvalues (diagonalizing matrix with repeated eigenvalues or a Jordan block of size greater than 1) have zero measure. Therefore generic Markov matrices have no repeated eigenvalues, and hence we are going to assume the eigenvalues to be distinct. In particular, we are assuming that M diagonalizes. Furthermore, since we want M to have real logarithms and have no repeated eigenvalues, we need the real eigenvalues to be positive.
The following theorem characterizes the embeddability of a CS Markov matrix with positive and distinct eigenvalues. Furthermore, the theorem guarantees that a CS Markov matrix is embeddable if and only if it admits a CS Markov generator. In particular, the characterization of the embeddability of a CS matrix is equivalent when restricting to rate matrices satisfying the symmetries imposed by the model (model embeddability) than when restricting to all possible rate matrices (embedding problem).
Theorem 1
Let M be a diagonalizable CS Markov matrix with positive and distinct eigenvalues defined as in (3.10). Let us define
where if and if Then any real logarithm of M is given by
where
with defined as in (3.3) and as in (3.8).
In particular, any real logarithm of M is also a CS matrix whose entries are given by:
Proof
Let us note that
Since we assume that the eigenvalues of M are distinct, according to Proposition 1, any logarithm of M can be written as
The last equation and the fact that S and are real matrices imply that Q will be real if and only if both and are real. Here is the upper block given in (3.7) and is the lower block defined in (3.9). By Lemma 3, being a real logarithm implies that and . Then being a real matrix, according to Lemma 4, implies that if , while if Therefore, the upper block is and the lower block will be completing the proof.
Now we are interested in knowing when the real logarithm of a CS Markov matrix is a rate matrix. Using the same notation as in Theorem 1 we get the following result.
Theorem 2
A diagonalizable CS Markov matrix M with distinct eigenvalues is embeddable if and only if the following conditions hold for if or for some if :
Proof
The logarithm of a CS Markov matrix will depend on whether or . In particular, it will depend on whether the eigenvalues and are real and positive or whether they are conjugated complex numbers.
If , then both and are real and . Hence, Moreover, Lemma 4 implies that and hence .
If , then and . Hence, and Moreover, since .
Thus, in both cases, Moreover, and are both non-positive. In particular, Theorem 1 together with Proposition 2 imply that a real logarithm of M is a rate matrix if and only if
Furthermore, the conditions comes from Lemma 3. The proof is now complete.
Remark 3
According to Theorem 2 the embeddability of a CS Markov matrix M with distinct positive eigenvalues can be decided by checking six inequalities depending on the entries of M. However, if M has non-real eigenvalues then one has to check infinitely many groups of inequalities, one for each value of . It is enough that one of those systems is consistent to guarantee that M is embeddable. Theorem 5.5 in Casanellas et al. (2023) provides boundaries for the values of k for which the corresponding inequalities may hold.
Let us take a look at the class of K3P matrices which is a special case of strand symmetric matrices. Indeed, for a K3P matrix , we have that
Suppose that a K3P-Markov matrix is K3P-embeddable, i.e. for some K3P-rate matrix Q. Recall that the eigenvalues of M are
In this case, we have that
In particular, we see that unless . Moreover,
The inequalities in Theorem 2 can be spelled out as follows:
These inequalities are equivalent to the K3P-embeddability criteria presented in (Roca-Lacostena and Fernández-Sánchez 2018b, Theorem 3.1) and (Ardiyansyah et al. 2021, Theorem 1). Moreover, they are also equivalent to the restriction to centrosymmetric-matrices of the embeddability criteria for Markov matrices with different eigenvalues given in (Casanellas et al. 2023, Theorem 1.1)
In the last part of this section, we discuss the rate identifiability problem for centrosymmetric matrices. If a centrosymmetric Markov matrix arises from a continuous-time model, then we want to determine its corresponding substitution rates. In other words, given an embeddable CS matrix, we want to know if we can uniquely identify its Markov generator.
It is worth noting that Markov matrices with repeated real eigenvalues may admit more than one Markov generator (e.g. examples 4.2 and 4.3 in (Casanellas et al. 2020a) show embeddable K2P matrices with more than one Markov generator). Nonetheless, this is not possible if the Markov matrix has distinct eigenvalues, because in this case its only possible real logarithm would be the principal logarithm (Culver 1966). As one considers less restrictions in a model, the measure of the set of matrices with repeated real eigenvalues decreases, eventually becoming a measure zero set. For example, this is the case within the K3P model, where both its submodels (the K2P model and the JC model) consist of matrices with repeated eigenvalues and have positive measure subsets of embeddable matrices with non-identifiable rates. However, when considering the whole set of K3P Markov matrices, the subset of embeddable matrices with more than one Markov generator has measure zero (see Chapter 4 in (Roca-Lacostena 2021)). Nevertheless, this behaviour only holds if the Markov matrices within the model have real eigenvalues.
Proposition 3
There is a positive measure subset of CS Markov matrices that are embeddable and whose rates are not identifiable. Moreover, all the Markov generators of the matrices in this set are also CS matrices.
Proof
Given
let us consider the following matrices
A straightforward computation shows that M is a CS Markov matrix and Q is a CS rate matrix. Moreover they both have non-zero entries. By applying the exponential series to Q, we get that . This means that M is embeddable and Q is a Markov generator of M.
Since Q is a rate matrix, so is Qt for any . Therefore, is an embeddable Markov matrix, because the exponential of any rate matrix is necessarily a Markov matrix. See (Pachter and Sturmfels 2005, Theorem 4.19) for more details. Moreover, we have that
so is a 2-block diagonal matrix. Hence, by Proposition 2 we have that is an embeddable strand symmetric Markov matrix for all .
Now, let us define . Note that Q and V diagonalize simultaneously via P and hence they commute. Therefore,
by the Baker-Campbell-Haussdorff formula. Moreover,
for all . Note that kV is a bounded matrix for any given k and hence, given t large enough, it holds that is a rate matrix for any m between 0 and k.
This shows that, for t large enough, is an embeddable CS Markov matrix with at least different CS Markov generators. Moreover, and all its generators have no null entries by construction and they can therefore be perturbed as in Theorem 3.3 in Casanellas et al. (2020b) to obtain a positive measure subset of embeddable CS Markov matrices that have CS Markov generators.
Remark 4
The perturbation presented in Theorem 3.3 in Casanellas et al. (2020b) consists of small changes on the real and complex parts of the eigenvalues and eigenvectors of M other than the eigenvector and its corresponding eigenvalue 1. If those changes are small enough then the resulting transition matrix and all its generators still satisfy the stochastic constraints of Markov/rate matrices.
Remark 5
Using the same notation as in the proposition above and given , let us define
Since is a CS rate matrix with no null entries, so is Q(C) for close enough to . Moreover, by construction we have that for all . Therefore, for we have that has uncountably many Markov generators (i.e. 2tQ(C) with C close to ) and all of them are CS matrices (Culver 1966, Corollary 1). It is worth noting that according to (Culver 1966, Corollary 1), if a matrix has uncountably many logarithms, then it necessarily has repeated real eigenvalues. Therefore, the subset of embeddable CS Markov matrices with uncountably many generators has measure zero within the set of all matrices.
Volumes of CS Markov matrices
In this section, we compute the relative volumes of embeddable CS Markov matrices within some meaningful subsets of Markov matrices. The aim of this section is to describe how large the different sets of matrices are compared to each other.
Let be the set of all CS Markov matrices. We use the following description
More explicitly, we identify the CS Markov matrix
with a point Let be the set of all CS Markov matrices having real positive eigenvalues, where
is the discriminant of the matrix as stated in Sect. 3. We have . More explicitly,
Let be the set of all embeddable CS Markov matrices with four distinct real positive eigenvalues. We have . Therefore, by Theorem 2,
Finally, we consider the following two biologically relevant subsets of . Let be the set of diagonally largest in column (DLC) Markov matrices, which is the subset of containing all CS Markov matrices such that the diagonal element is the largest element in each column. These matrices are related to matrix parameter identifiability in phylogenetics (Chang 1996). Secondly, we let be the set of diagonally dominant (DD) Markov matrices, which is the subset of matrices containing all CS Markov matrices such that in each row the diagonal element is at least the sum of all the other elements in the row. Biologically, the subspace consists of matrices with probability of not mutating at least as large as the probability of mutating. If a diagonally dominant matrix is embeddable, it has an identifiable rate matrix (Cuthbert 1972; James 1973). By the definition of each set, we have the inclusion .
Remark 6
The sets , , , that we consider in this section are all subsets of the set of all CS Markov matrices, but we can use the same definition to refer to the equivalent subsets of CS Markov matrices. Therefore, we will use the same notation , , , to refer to the equivalent subsets of the set of CS Markov matrices without confusion in the following sections.
In the rest of this section, the number v(A) denotes the Euclidean volume of the set A. By definition, , and are polytopes, since they are defined by the linear inequalities in . Hence, we can use Polymake (Gawrilow and Joswig 2000) to compute their exact volumes and obtain that
Hence, we see that and constitute roughly only and of , respectively.
On the other hand, we will estimate the volume of the sets using the hit-and-miss Monte Carlo integration method (Hammersley 2013) with sufficiently many sample points in Mathematica (Inc. 2022). Theoretically, Theorem 2 enables us to compute the exact volume of these relevant sets. For example in the case of K3P matrices, such exact computation of volumes has been feasible in Roca-Lacostena and Fernández-Sánchez (2018b). However, while for the K3P matrices, the embeddability criterion is given by three quadratic polynomial inequalities, in the case of CS matrices the presence of nonlinear and nonpolynomial constraints imposed on each set, makes the exact computation of the volume of these sets intractable. Therefore, we need to approximate the volume of these sets. Given a subset , the volume estimate of v(A) computed using the hit-and-miss Monte Carlo integration method with n sample points is given by the number of points belonging to A out of n sample points. For computational purposes, in the formula of and , we use the fact that
All codes for the computations implemented Mathematica and Polymake can be found at the following address: https://github.com/ardiyam1/Embeddability-and-rate-identifiability-of-centrosymmetric-matrices.
The results of these estimations using the hit-and-miss Monte Carlo integration implemented in Mathematica with n sample points are presented in Table 1, while Table 2 provides an estimated volume ratio between relevant subsets of centrosymmetric Markov matrices using again the hit-and-miss Monte Carlo integration with n sample points. In Table 1, we firstly generate n centrosymmetric matrices whose off-diagonal entries were sampled uniformly in [0, 1] and forced the rows of the matrix to sum to one. Out of these n matrices, we test how many of them are actually Markov matrices (i.e. the diagonal entries are non-negative) and then out of these how many have positive eigenvalues. In particular, for sample points containing 277628 centrosymmetric Markov matrices, Table 2 suggests that there are approximately of centrosymmetric Markov matrices with distinct positive eigenvalues that are embeddable. Moreover, we can see that for , out of all embeddable centrosymmetric Markov matrices with distinct positive eigenvalues, almost all are diagonally largest in column, while only are diagonally dominant.
Table 1.
Number of samples in the sets
| n | ||||
|---|---|---|---|---|
| Samples in | 280 | 2767 | 27829 | 277628 |
| Samples | 23 | 192 | 1999 | 20601 |
| Samples in | 3 | 34 | 359 | 3511 |
| Samples in | 19 | 154 | 1541 | 15830 |
| Samples in | 3 | 31 | 262 | 2889 |
| Samples in | 3 | 34 | 357 | 3503 |
| Samples in | 1 | 15 | 105 | 1011 |
using hit-and-miss methods and Theorem 2.
Table 2.
Relative volumes ratio between the relevant subsets obtained using hit-and-miss method and Theorem 2. The volumes were estimated as the quotient of the sample sizes in Table 1
| n | ||||
|---|---|---|---|---|
| 0.130435 | 0.177083 | 0.17959 | 0.170429 | |
| 0.157895 | 0.220779 | 0.231668 | 0.221289 | |
| 0.130435 | 0.177083 | 0.178589 | 0.17004 | |
| 1 | 1 | 0.994429 | 0.997721 | |
| 0.333333 | 0.483871 | 0.400763 | 0.349948 | |
| 0.0434783 | 0.078125 | 0.052563 | 0.0490753 | |
| 0.333333 | 0.441176 | 0.292479 | 0.287952 |
An alternative approach for approximating the number of embeddable matrices within the model is to use Algorithm 5.8 in Casanellas et al. (2023) to test the embeddability of the sample points. Tables 4 and 5 below are analogous to Tables 1 and 2, but Table 4 was obtained using the sampling method in (Roca-Lacostena 2021, Appendix A), while using either Algorithm 5.8 in Casanellas et al. (2023) or the inequalities in Theorem 2 yields identical results which are provided in Tables 4 and 5.
Table 4.
Number of samples in obtained by applying either Theorem 2 or the results in Casanellas et al. (2023) on the sample set in Table 3
| Samples in | 1877 | 18663 | 185357 | 1862413 |
|---|---|---|---|---|
| Samples in | 1869 | 18586 | 184555 | 1854592 |
| Samples in | 516 | 5164 | 50058 | 504304 |
Table 5.
Relative volumes ratio between the relevant subsets obtained using hit-and-miss method and either Algorithm 5.8 in Casanellas et al. (2023) or Theorem 2. The volumes were estimated as the quotient of the sample sizes in Tables 3 and 4
| n | ||||
|---|---|---|---|---|
| 0.1877 | 0.18663 | 0.185357 | 0.1862413 | |
| 0.2191 | 0.2175 | 0.2159 | 0.2169 | |
| 0.1869 | 0.18586 | 0.184555 | 0.1854592 | |
| 0.9957 | 0.9959 | 0.99567 | 0.99580 | |
| 0.3524 | 0.3552 | 0.3463 | 0.3481 | |
| 0.0516 | 0.05164 | 0.050058 | 0.0504 | |
| 0.2749 | 0.2767 | 0.2701 | 0.2708 |
We used the python implementation of Algorithm 5.8 in Casanellas et al. (2023) provided in (Roca-Lacostena 2021, Appendix A) and modified it to sample on the set of CS Markov matrices with positive eigenvalues. The original sampling method used in (Roca-Lacostena 2021, Appendix A) consisted of sampling uniformly on the set of centrosymmetric-Markov matrices until we obtained n samples (or as many samples as we require) with positive eigenvalues.
Despite the fact that Theorem 2 and Algorithm 5.8 in Casanellas et al. (2023) were originally implemented using different programming languages (Wolfram Mathematica and Python respectively) and were tested with different sample sets, the results obtained are quite similar as illustrated by Tables 2 and 5. In fact, when we apply both Algorithm 5.8 in Casanellas et al. (2023) and Theorem 2 on the same sample set in Table 3, we obtain identical results which are displayed in Tables 4 and 5.
Table 3.
Number of samples in obtained by using the sampling method in (Roca-Lacostena 2021, Appendix A)
| Samples in | ||||
|---|---|---|---|---|
| Samples in | 8531 | 85446 | 854709 | 8549100 |
| Samples in | 1464 | 14538 | 144546 | 1448720 |
It is worth noting that the embeddability criteria given in Theorem 2 use inequalities depending on the entries of the matrix, whereas Algorithm 5.8 in Casanellas et al. (2023) relies on the computation of its principal logarithm and its eigenvalues and eigenvector, which may cause numerical issues when working with matrices with determinant close to 0. Moreover, the computation of logarithms can be computationally expensive. As a consequence, the algorithm implementing the criterion for embeddability arising from Theorem 2 is faster. Table 6 shows the running times for the implementation of both embeddability criteria used to obtain Table 5.
Table 6.
Running times for the Python implementation of the embeddability criterion arising from Theorem 2 and from Algorithm 5.8 in Casanellas et al. (2023). The simulations were run using a computer with 8GB of memory
| Sampling time | 12.5s | 121.5s (2 min) | 1222s (20min) | 12141.8s (3h 22min) |
| Embedding criteria (Theorem 2) | 28.3s | 273.2s (4min 30s) | 2703s (45min) | 27413s (7h 37min) |
| Embedding criteria (Algorithm 5.8) | 84.2s | 840.5s (15 min) | 8358 (2h 19min) | 83786s (23h 16min) |
The Python implementation of Algorithm 5.8 in Casanellas et al. (2023) provided in (Roca-Lacostena 2021, Appendix A) can also be used to test the embeddability of any CS Markov matrix (including those with non-real eigenvalues) without modifying the embeddability criteria. To do so, it is enough to apply the algorithm to a set of Markov matrices with different eigenvalues sampled uniformly from the set of all CS Markov matrix. As hinted in Remark 3, this would also be possible using the embedability criterion in Theorem 2 together with the boundaries for k provided in (Casanellas et al. 2023, Theorem 5.5). Table 7 shows the results obtained when applying Algorithm 5.8 in Casanellas et al. (2023) to a set of CS Markov matrices sampled uniformly.
Table 7.
Embeddable matrices within CS Markov matrices and its intersection with DLC matrices and DD matrices
| Samples | Embeddable samples | Proportion of embeddable | |
|---|---|---|---|
| 173455 | 0.0173455 | ||
| 1021195 | 172380 | 0.1688022 | |
| 156637 | 49471 | 0.3158321 |
As most DLC and DD matrices have positive eigenvalues, the proportion of embeddable matrices within these subsets is almost the same when admitting matrices with non-positive eigenvalues (as in Table 7 instead of only considering matrices with positive eigenvalues as we did in Tables 2 and 5. On the other hand, the proportion of embeddable CS matrices is much smaller in this case.
Centrosymmetric matrices and generalized Fourier transformation
In Sects. 3 and 4 we have seen the embeddability criteria for centrosymmetric Markov matrices and the volume of their relevant subsets. In this section, we are extending this framework to larger matrices. The importance of this extension is relevant to the goal of synthetic biology which aims to expand the genetic alphabet. For several decades, scientists have been cultivating ways to create novel forms of life with basic biochemical components and properties far removed from anything found in nature. In particular, they are working to expand the number of amino acids which is only possible if they are able to expand the genetic alphabet (see for example (Hoshika et al. 2019)).
Properties of centrosymmetric matrices
For a fixed , let denote the set of all centrosymmetric matrices of order n. Moreover, let and denote the set of all centrosymmetric Markov and rate matrices of order n, respectively. As a subspace of the set of all real matrices, for n even, dim while for n odd, dim. We will now mention some geometric properties of the sets and Furthermore, for any real number x, and denote the floor and the ceiling function of x, respectively.
Proposition 4
For n even, is a Cartesian product of standard -simplices and its volume is For n odd, is a Cartesian product of standard -simplices and the -simplex with vertices , where is the i-th standard unit vector in . Hence, the volume of is .
For n even, and for n odd, .
Proof
Here we consider the following identification for an centrosymmetric matrix M. For n even, M can be thought as a point where the point corresponds to the i-th row of M. Similarly, for n odd, we identify M as a point in . Since M is a Markov matrix, under this identification, each point lies in some simplices. Therefore, is a Cartesian product of some simplices. For n even, these simplices are the standard -dimensional simplex:
| 5.1 |
For n odd and , the point belongs to standard -simplex above and the point belongs to the simplex
| 5.2 |
We now compute the volume of . Let us recall the fact that the volume of the Cartesian product of spaces is equal to the product of volumes of each factor space if the volume of each factor space is bounded. Moreover, the -dimensional volume of the standard simplex in Eq. (5.1) in is For n even, the statement follows immediately. For n odd, we use the fact that the -dimensional volume of the simplex in Eq. (5.2) is We refer the reader to Stein (1966) for an introductory text on the volume of simplices.
For the second statement, we use the fact that if Q is a rate matrix, then where for .
In the rest of this section, let be the anti-diagonal matrix, i.e. the (i, j)-entries are one if and zero otherwise. The following proposition provides some properties of the matrix that can be checked easily.
Proposition 5
Let . Then
A is a centrosymmetric matrix if only if
In Sect. 3, we have seen that CS matrices can be block-diagonalized through the matrix S. Now we will present a construction of generalized Fourier matrices to block-diagonalize any centrosymmetric matrices. Let us consider the following recursive construction of the matrix :
| 5.3 |
Proposition 6
For each natural number , is invertible and its inverse is given by
Proof
The proposition easily follows from the definition of . Indeed, we have
The following proposition provides another block decomposition of the matrix and its inverse.
Proposition 7
Let
For n even, , while for n odd,
Using these block partitions, for n even, while for n odd.
Proof
The proof follows from induction on n and the fact that .
We will call a vector symmetric if for every , i.e. Moreover, we call a vector anti-symmetric if for every , i.e. The following technical proposition will be used in what follows in order to simplify a centrosymmetric matrix.
Proposition 8
Let . Let be a symmetric vector and be an anti-symmetric vector.
The last entries of and are zero. Similarly, the last entries of and are zero.
The first entries of and are zero. Similarly, the first entries of and are zero.
Then the sum of the entries of and is the sum of the entries of v.
Then the sum of the entries of and is the sum of the first entries of v.
Proof
We will only prove the first part of item (1) in the proposition using mathematical induction on n. The base case for can be easily obtained. Suppose now that the proposition holds for all Let be a symmetric element. Then is also symmetric. By direct computation we obtain
The last entries of are zero. Thus, the last entries of are zero as well. The proof of the other statements can be obtained analogously using induction. In particular, let us note that the proof given for item (1) directly implies item (3).
For a fixed number n, let us define the following map:
For we have seen that if A is a CS matrix, then is a block-diagonal matrix where each block is of size and is given by and . Moreover, the upper block is a Markov matrix. The following lemma provides a generalization to these results.
Lemma 5
Let Given an CS matrix A, is the following block-diagonal matrix
where is a matrix of size . Furthermore, if A is a Markov (rate) matrix, then is also a Markov (rate) matrix.
Proof
First suppose that n is even. By (Cantoni and Butler 1976, Lemma 2), we can partition A into the following block matrices:
where and are of size By Proposition 7, we have
Choose . Now suppose that A is a Markov matrix. This means that each row of A sums to 1 and A has non-negative entries. Therefore, for , we have
and for ,
Now we consider the case when n is odd. Again by (Cantoni and Butler 1976, Lemma 2), we can partition A into the following block matrices:
where , and By Proposition 7, we have
In this case, choose Suppose that A is a Markov matrix. Since each row of A sums to 1, we have
and for ,
From the fact that the entries of A are non-negative, for , we obtain that
Therefore, all entries of sum to 1 and are non-negative meaning that is a Markov matrix as well. We can proceed similarly for the case when A is a rate matrix.
Lemma 6
For any natural number n, let . Suppose that is a block diagonal matrix. Then
is a CS matrix.
- is a Markov matrix if and only if is a Markov matrix and for any
- is a rate matrix if and only if is a rate matrix and for any such that for , and for ,
Proof
We will only prove the lemma for n even. Similar arguments will work for n odd as well. By Proposition 7,
Since and , then by (Cantoni and Butler 1976, Lemma 2), is centrosymmetric which proves (1). For ,
The above equality means that for , the i-th row sum of and coincide. This implies that if is a Markov (rate) matrix, then is a Markov (rate) matrix as well. Additionally, note that
Hence, (2) and (3) will follow immediately.
Logarithms of centrosymmetric matrices
For the special structure encoded by the centrosymmetric matrices, one may ask whether they have logarithms which are also centrosymmetric. In this section, we provide some answers to this question.
Theorem 3
Let be a CS matrix. Then A has a CS logarithm if and only if both the upper block matrix and the lower block matrix in Lemma 5 admit a logarithm.
Proof
Suppose that A has a centrosymmetric logarithm Q. By Lemma 5, and Then implies that and Hence, and admit a logarithm. Conversely, suppose that and admit a logarithm and , respectively. Then the matrix is a logarithm of the matrix . By Lemma 6, the matrix is a centrosymmetric logarithm of A.
Proposition 9
Let be a CS matrix. If A is invertible, then it has infinitely many CS logarithms.
Proof
The assumptions imply that the matrices and in Lemma 5 are invertible. By (Higham 2008, Theorem 1.28), each and has infinitely many logarithms. Hence, Theorem 3 implies that A has infinitely many centrosymmetric logarithms.
Proposition 10
Let be a CS matrix such that Log(A) is well-defined. Then Log(A) is again centrosymmetric.
Proof
Let us suppose that Log(A) is not centrosymmetric matrix. Define the matrix . Then since Log(A) is not centrosymmetric. It is also clear that . Moreover, since , the matrices Log(A) and Q have the same eigenvalues. Therefore, Q is also a principal logarithm of A, a contradiction to the uniqueness of principal logarithm. Hence, Log(A) must be centrosymmetric.
The following theorem characterizes the logarithms of any invertible CS Markov matrices.
Theorem 4
Let be an invertible CS Markov matrix. Let where is a Jordan form of the upper block matrix in Lemma 5. Similarly, let where is a Jordan form of the lower block matrix in Lemma 5. Then A has a countable infinitely many logarithms given by
where
and denotes a logarithm of . In particular, these logarithms of A are primary functions of A.
Proof
The theorem follows immediately from (Higham 2008, Theorem 1.28).
For the definition of primary function of a matrix, we refer the reader to Higham (2008). The above theorem says that the logarithms of a nonsingular centrosymmetric matrix contains a countable infinitely many primary logarithms and they are centrosymmetric matrices as well.
Finally, we will present a necessary condition for embeddability of CS Markov matrices in higher dimensions.
Lemma 7
Let . Suppose that is an embeddable CS Markov matrix of size with a CS logarithm. Then for n even,
while for n odd,
Proof
Since A is an embeddable matrix with CS logarithm, we write for some CS rate matrix Q, and then
By Lemma 5, for the centrosymmetric matrices A, Q, we have and where is a Markov matrix and is a rate matrix of size Therefore, If are the eigenvalues, perhaps not distinct, of then the eigenvalues of are . Since one of ’s is zero, then the trace of which is the sum of its eigenvalues is equal to
We now need to show that trace of has the form written in the lemma. Suppose that n is even. By the proof of Lemma 5, then
The proof for odd n can be obtained similarly.
Let be the subset containing all centrosymmetric-embeddable Markov matrices. We want to obtain an upper bound of the volume of using Lemma 7. Let be the subset containing all centrosymmetric Markov matrices such that after applying the generalized Fourier transformation, the trace of the upper block matrix is greater than 1. The previous lemma implies that and hence, . Moreover, the upper bound is easy to compute as is a polytope and for some values of n, these volumes are presented in Table 8. We see from Table 8, there are at most of matrices in that are centrosymmetically-embeddable and hence this upper bound is not good. For , approximately, there are at most in that are centrosymmetrically-embeddable but for , this upper bound gives a better proportion, which is approximately .
Table 8.
The exact volume computed using Polymake
| Dimension of | |||
|---|---|---|---|
| 6 | |||
| 10 | |||
| 15 |
Embeddability of centrosymmetric matrices
Throughout this section we shall consider A to be a centrosymmetric Markov matrix with distinct eigenvalues. In particular, the matrices considered in this section are diagonalizable and are a dense subset of all centrosymetric Markov matrices. Note that this notation differs from the notation for Markov matrices used in previous sections in order to make it consistent with the notation used in the results presented for generic centrosymmetric matrices.
In the previous section, we showed that F(A) is a block-diagonal real matrix composed of two blocks denoted by and . Since both and have real entries, each of these matrices has at most one conjugate pair of eigenvalues. Adapting the notation introduced in Theorem 4 to diagonalizable matrices we have such that and with and . Moreover, we can assume that without loss of generality (this can be achieved by permuting the second and third columns of if necessary). For ease of reading, we will define as , where is the matrix used to obtain the Fourier transform F(A) and was introduced in Sect. (5.3).
Next we give a criterion for the embeddability of A for each of the following cases: 
Proposition 11
If a cetrosymmetric Markov matrix A does not belong to any of the cases in Table 6.1, then it is not embeddable.
Proof
If A satifies the hypothesis of the proposition then either it has a null eigenvalue or it has a simple negative eigenvalue. In the former case A is a singular matrix and hence it has no logarithm. If A had a simple negative eigenvalue, then all its logarithms would have a non-real eigenvalue whose complementary pair is not an eigenvalue of A (otherwise M would have a repeated eigenvalues). Therefore, A has no real logarithm.
Remark 7
All the results in this section can be adapted to centrosymmetric Markov matrices by not considering the eigenvalue and modifying the forthcoming definitions of the matrices and V accordingly (i.e. removing the fourth row and column in the corresponding diagonal matrix). In addition, these results still hold if the eigenvalue 1 of the Markov matrix has multiplicity 2.
Case 1
The results for this case are not restricted to centrosymmetric matrices but can be applied to decide the embeddability of any suitable Markov matrix.
Proposition 12
If all the eigenvalues of a Markov matrix A are distinct and positive, then A is embeddable if and only if Log(A) is a rate matrix.
Proof
If A has distinct real eigenvalues then it has only one real logarithm, which is Log(A) (see (Culver 1966)).
Case 2
In this case A has exactly one conjugate pair of complex eigenvalues and we obtain the following criterion by adapting Corollary 5.6 in Casanellas et al. (2023) to our framework:
Proposition 13
Given the matrix define:
and set Then,
A is embeddable if and only if and .
the set of Markov generators for A is .
Proof
The proof of this theorem is analogous to the proof of Theorem 5.5 in Casanellas et al. (2020a) but considering the matrix V as defined here. According to Proposition 1, any Markov generator of A is of the form
Such a logarithm can be rewritten as . Using this, we will prove that is a rate matrix if and only if and .
Suppose that there exists such that is a rate matrix. Hence, for all . For , we have:
for all such that . This means that
for all such that . This means that .
for all such that . This means that .
Conversely, suppose that and and that there is such that . We want to check that is a rate matrix. According to Proposition 1, each row of sums to 0. Moreover, for , we have:
if , then . Since , .
if , then
if , then
The proof is now complete.
Case 3
As in Case 2, A has exactly one conjugate pair of eigenvalues and hence its embeddability (and all its generators) can be determined by using Proposition 13 but defining the matrix V as . However in Case 3 the conjugate pair of eigenvalues lie in which is a Markov matrix. This allows us to use the results regarding the embeddability of Markov matrices to obtain an alternative criterion to test the embeddability of A. To this end we define
| 6.2 |
where .
Proposition 14
The matrix A is embeddable if and only if Log(A) or are rate matrices.
Proof
Note that so one of the implications is immediate to prove. To prove the other implication, we assume that A is embeddable and let Q be a Markov generator for it. Proposition 1 yields that
for some integers . Therefore, where and are real logarithms of and respectively.
Since is a real matrix with distinct positive eigenvalues, its only real logarithm is its principal logarithm. This implies that (so that ).
Now, recall that is a Markov matrix (see Lemma 5). Using Proposition 1 again, we obtain that is a rate matrix, thus is embeddable. To conclude the proof it is enough to recall Theorem 4 in James (1973), which yields that is embeddable if and only if or is a rate matrix.
Case 4
In this case, the solution to the embedding problem can be obtained as a byproduct of the results for the previous cases:
Proposition 15
Let denote the principal logarithm of A and denote the matrix in (6.2). Given the matrix and define:
and set Then,
A is embeddable if and only if and for or .
- If A is embeddable, then at least one of its Markov generator can be written as
with and such that .
Proof
The matrix A is embeddable if and only if it admits a Markov generator. According to Proposition 1, if such a generator Q exists then it can be written as for some . Therefore, Lemma 5 implies that for some matrices and . Moreover, where and are real logarithms of and respectively.
As shown in the proof of Proposition 14, is actually a Markov matrix and is a Markov generator for it (see also Lemma 5). Moreover, by Theorem 4 in James (1973), is embeddable if and only if or are rate matrices. This implies that is a rate matrix if and only if or are rate matrices. To conclude the proof we proceed as in the proof of Proposition 13. Indeed, note that for , . Using this, it is immediate to check that is a rate matrix if and only if and .
Discussion
The central symmetry is motivated by the complementarity between both strands of the DNA. When a nucleotide substitution occurs in one strand, there is also a substitution between the corresponding complementary nucleotides on the other strand. Therefore, working with centrosymmetric Markov matrices is the most general approach when considering both DNA strands.
In this paper, we have discussed the embedding problem for centrosymmetric Markov matrices. In Theorem 2, we have obtained a characterization of the embeddabilty of centrosymmetric Markov matrices which are exactly the strand symmetric Markov matrices. In particular, we have also shown that if a CS Markov matrix is embeddable, then any of its Markov generators is also a CS matrix. Furthermore, In Sect. 6, we have discussed the embeddability criteria for larger centrosymmetric matrices.
As a consequence of the characterization of Theorem 2, we have been able to compute and compare the volume of the embeddable CS Markov matrices within some subspaces of CS Markov matrices. These volume comparisons can be seen in Table 2 and Table 7. For larger matrices, using the results in Sect. 6, we have estimated the proportion of embeddable matrices within the set of all centrosymmetric Markov matrices and within the subsets of DLC and DD matrices. This is summarized in Table 9 below. The computations were repeated several times obtaining results with small differences in the values but the same order of magnitude and starting digits.
Table 9.
Relative volume of embeddable matrices within relevant subsets of centrosymmetric Markov matrices. The results were obtained using the hit-and-miss Monte Carlo integration with sample points
| Set | Sample points | Embeddable sample points | Rel. vol. of embeddable matrices |
|---|---|---|---|
| 1370 | 0.0000137 | ||
| 1034607 | 1362 | 0.0013164 | |
| 3048 | 84 | 0.0275590 |
As we have seen in Sect. 3 and 6, we have only considered in detail the embeddability of CS Markov matrices of size and . We expect that the proportion of the embeddable CS Markov matrices within the subset of Markov matrices in larger dimension tends to zero as n grows larger as indicated by Tables 2, 7, 8, and 9.
These results together with the results obtained for the strand symmetric model (see Table 7) indicate that restricting to homogeneous Markov processes in continuous-time is a very strong restriction because non-embeddable matrices are discarded and their proportion is much larger than that of embeddable matrices. For instance, in the case exactly of the matrices are discarded (Ardiyansyah et al. 2021, Table 5), while in the case of matrices up to of the matrices are discarded (see Table 7) and in the case of matrices the amount of discarded matrices is about as indicated in Table 9. However, when restricting to subsets of Markov matrices which are mathematically more meaningful in biological terms, such as DD or DLC matrices, the proportion of embeddable matrices is much higher so that we are discarding less matrices (e.g. for DD we discard of matrices and of matrices). This is not to say that it makes no sense to use continuous-time models but to highlight that one should take the above restrictions into consideration when working with these models. Conversely, when working with the whole set of Markov matrices one has to be aware that they might end up considering lots of non-meaningful matrices.
Acknowledgements
Dimitra Kosta was partially supported by a Royal Society Dorothy Hodgkin Research Fellowship DHFR1201246. Jordi Roca-Lacostena was partially funded by Secretaria d’Universitats i Recerca de la Generalitat de Catalunya (AGAUR 2018FI_B_0094). Muhammad Ardiyansyah is partially supported by the Academy of Finland Grant No. 323416.
Funding Information
Open Access funding provided by Aalto University.
Declarations
Conflict of interest
The authors declare no conflict of interest.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Muhammad Ardiyansyah, Email: muhammad.ardiyansyah@aalto.fi.
Dimitra Kosta, Email: D.Kosta@ed.ac.uk.
Jordi Roca-Lacostena, Email: jordi.roca.lacostena@upc.edu.
References
- Aitken AC (2017) Determinants and matrices. Read Books Ltd
- Ardiyansyah M, Kosta D, Kubjas K. The model-specific Markov embedding problem for symmetric group-based models. J Math Biol. 2021;83(3):1–26. doi: 10.1007/s00285-021-01656-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baake M, Sumner J. Notes on markov embedding. Linear Algebra Appl. 2020;594:262–299. doi: 10.1016/j.laa.2020.02.016. [DOI] [Google Scholar]
- Bain J, Switzer C, Chamberlin R, Benner SA. Ribosome-mediated incorporation of a non-standard amino acid into a peptide through expansion of the genetic code. Nature. 1992;356(6369):537–539. doi: 10.1038/356537a0. [DOI] [PubMed] [Google Scholar]
- Benner SA, Sismour AM. Synthetic Biol. Nat Rev Genet. 2005;6(7):533–543. doi: 10.1038/nrg1637. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cantoni A, Butler P. Eigenvalues and eigenvectors of symmetric centrosymmetric matrices. Linear Algebra Appl. 1976;13(3):275–288. doi: 10.1016/0024-3795(76)90101-4. [DOI] [Google Scholar]
- Carette P. Characterizations of embeddable 33 stochastic matrices with a negative eigenvalue. New York J Math. 1995;1:120–129. [Google Scholar]
- Casanellas M, Kedzierska AM. Generating Markov evolutionary matrices for a given branch length. Linear Algebra Appl. 2013;438(5):2484–2499. doi: 10.1016/j.laa.2012.11.009. [DOI] [Google Scholar]
- Casanellas M, Sullivant S (2005) The strand symmetric model. In: Pachter L, Sturmfels B (eds) Algebraic statistics for computational biology. Cambridge University Press, New York
- Casanellas M, Fernández-Sánchez J, Roca-Lacostena J. Embeddability and rate identifiability of Kimura 2-parameter matrices. J Math Biol. 2020;80(4):995–1019. doi: 10.1007/s00285-019-01446-0. [DOI] [PubMed] [Google Scholar]
- Casanellas M, Fernández-Sánchez J, Roca-Lacostena J (2020b) An open set of 4 4 embeddable matrices whose principal logarithm is not a markov generator. Linear and Multilinear Algebra pp 1–12
- Casanellas M, Fernández-Sánchez J, Roca-Lacostena J. The embedding problem for Markov matrices. Publicacions Matemàtiques. 2023;67:411–445. doi: 10.5565/PUBLMAT6712308. [DOI] [Google Scholar]
- Chang JT. Full reconstruction of Markov models on evolutionary trees: identifiability and consistency. Math Biosci. 1996;137(1):51–73. doi: 10.1016/S0025-5564(96)00075-2. [DOI] [PubMed] [Google Scholar]
- Chen Y, Chen J. On the imbedding problem for three-state time homogeneous Markov chains with coinciding negative eigenvalues. J Theor Probab. 2011;24:928–938. doi: 10.1007/s10959-010-0316-5. [DOI] [Google Scholar]
- Culver WJ. On the existence and uniqueness of the real logarithm of a matrix. Proceed Am Math Soc. 1966;17:1146–1151. doi: 10.1090/S0002-9939-1966-0202740-6. [DOI] [Google Scholar]
- Cuthbert JR. On uniqueness of the logarithm for Markov semi-groups. J Lond Math Soc. 1972;2(4):623–630. doi: 10.1112/jlms/s2-4.4.623. [DOI] [Google Scholar]
- Davies E, et al. Embeddable Markov matrices. Electron J Probab. 2010;15:1474–1486. doi: 10.1214/EJP.v15-733. [DOI] [Google Scholar]
- Elfving G (1937) Zur theorie der Markoffschen ketten. Acta Societatis Scientiarum FennicæNova Series A 2(8):17 pages
- Fuglede B. On the imbedding problem for stochastic and doubly stochastic matrices. Probab Theory Relat Fields. 1988;80:241–260. doi: 10.1007/BF00356104. [DOI] [Google Scholar]
- Gawrilow E, Joswig M (2000) Polymake: a framework for analyzing convex polytopes. In: Polytopes-combinatorics and computation, Springer, pp 43–73
- Goodman GS. An intrinsic time for non-stationary finite Markov chains. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete. 1970;16:165–180. doi: 10.1007/BF00534594. [DOI] [Google Scholar]
- Hammersley J (2013) Monte carlo methods. Springer Science & Business Media
- Higham NJ (2008) Functions of matrices: Theory and computation, vol 104. SIAM
- Hoshika S, Leal NA, Kim MJ, Kim MS, Karalkar NB, Kim HJ, Bates AM, Watkins NE, SantaLucia HA, Meyer AJ, et al. Hachimoji DNA and RNA: a genetic system with eight building blocks. Science. 2019;363(6429):884–887. doi: 10.1126/science.aat0971. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Inc WR (2022) Mathematica, Version 13.1 https://www.wolfram.com/mathematica
- Iosifescu M (2014) Finite Markov processes and their applications. Courier Corporation
- James CR. The logarithm function for finite-state Markov semi-groups. J Lond Math Soc. 1973;2(3):524–532. [Google Scholar]
- Jia C. A solution to the reversible embedding problem for finite markov chains. Statist Probab Lett. 2016;116:122–130. doi: 10.1016/j.spl.2016.04.020. [DOI] [Google Scholar]
- Johansen S (1974) Some results on the imbedding problem for finite Markov chains. J London Math Soc s2-8(2):345–351
- Kimura M (1957) Some problems of stochastic processes in genetics. Ann Math Statistics pp 882–901
- Kingman JFC. The imbedding problem for finite Markov chains. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete. 1962;1(1):14–24. doi: 10.1007/BF00531768. [DOI] [Google Scholar]
- Leal NA, Kim HJ, Hoshika S, Kim MJ, Carrigan MA, Benner SA. Transcription, reverse transcription, and analysis of RNA containing artificial genetic components. ACS Synth Biol. 2015;4(4):407–413. doi: 10.1021/sb500268n. [DOI] [PubMed] [Google Scholar]
- Malyshev DA, Dhami K, Quach HT, Lavergne T, Ordoukhanian P, Torkamani A, Romesberg FE (2012) Efficient and sequence-independent replication of DNA containing a third base pair establishes a functional six-letter genetic alphabet. Proceedings of the National Academy of Sciences 109(30):12,005–12,010 [DOI] [PMC free article] [PubMed]
- Pachter L, Sturmfels B. Algebraic statistics for computational biology, UK: Cambridge University Press; 2005. [Google Scholar]
- Roca-Lacostena J (2021) The embedding problem for Markov matrices. PhD thesis, Universitat Politècnica de Catalunya
- Roca-Lacostena J, Fernández-Sánchez J. Embeddability of Kimura 3ST Markov matrices. J Theor Biol. 2018;445:128–135. doi: 10.1016/j.jtbi.2018.02.005. [DOI] [PubMed] [Google Scholar]
- Roca-Lacostena J, Fernández-Sánchez J. Embeddability of Kimura 3st markov matrices. J Theor Biol. 2018;445:128–135. doi: 10.1016/j.jtbi.2018.02.005. [DOI] [PubMed] [Google Scholar]
- Runnenberg JT. On Elfving’s problem of imbedding a time-discrete markov chain in a time-continuous one for finitely many states. Proceedings of the KNAW - Series A, Mathematical Sciences. 1962;65:536–541. [Google Scholar]
- Schensted IV. Appendix model of subnuclear segregation in the macronucleus of ciliates. Am Nat. 1958;92(864):161–170. doi: 10.1086/282023. [DOI] [Google Scholar]
- Sismour AM, Lutz S, Park JH, Lutz MJ, Boyer PL, Hughes SH, Benner SA. Pcr amplification of DNA containing non-standard base pairs by variants of reverse transcriptase from human immunodeficiency virus-1. Nucleic Acids Res. 2004;32(2):728–735. doi: 10.1093/nar/gkh241. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stein P. A note on the volume of a simplex. Am Math Mon. 1966;73(3):299–301. doi: 10.2307/2315353. [DOI] [Google Scholar]
- Weaver JR. Centrosymmetric (cross-symmetric) matrices, their basic properties, eigenvalues, and eigenvectors. Am Math Mon. 1985;92(10):711–717. doi: 10.1080/00029890.1985.11971719. [DOI] [Google Scholar]
- Yang Z, Hutter D, Sheng P, Sismour A, Benner S. Artificially expanded genetic information system: a new base pair with an alternative hydrogen bonding pattern. Nucleic Acids Res. 2006;34(21):6095–101. doi: 10.1093/nar/gkl633. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang Z, Sismour AM, Sheng P, Puskar NL, Benner SA. Enzymatic incorporation of a third nucleobase pair. Nucleic Acids Res. 2007;35(13):4238–4249. doi: 10.1093/nar/gkm395. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yang Z, Chen F, Alvarado JB, Benner SA (2011) Amplification, mutation, and sequencing of a six-letter synthetic genetic system. J Am Chem Soc 133(38):15,105–15,112 [DOI] [PMC free article] [PubMed]
- Yap VB, Pachter L. Identification of evolutionary hotspots in the rodent genomes. Genome Res. 2004;14(4):574–579. doi: 10.1101/gr.1967904. [DOI] [PMC free article] [PubMed] [Google Scholar]
