Abstract
Cancer progression can be described by continuous-time Markov chains whose state space grows exponentially in the number of somatic mutations. The age of a tumor at diagnosis is typically unknown. Therefore, the quantity of interest is the time-marginal distribution over all possible genotypes of tumors, defined as the transient distribution integrated over an exponentially distributed observation time. It can be obtained as the solution of a large linear system. However, the sheer size of this system renders classical solvers infeasible. We consider Markov chains whose transition rates are separable functions, allowing for an efficient low-rank tensor representation of the linear system’s operator. Thus we can reduce the computational complexity from exponential to linear. We derive a convergent iterative method using low-rank formats whose result satisfies the normalization constraint of a distribution. We also perform numerical experiments illustrating that the marginal distribution is well approximated with low rank.
Keywords: Transient distribution, Stochastic Automata Networks, Mutual Hazard Networks, Hierarchical Tucker format
Introduction
The dynamics of cancer can be studied using tumor progression models, cf. (Beerenwinkel et al. 2014). These models describe the evolving genotype of a tumor as a continuous-time Markov chain. A model includes d genomic loci that may or may not be mutated. It starts out with all mutations absent and then progressively accumulates mutations (or other genomic events such as copy number alterations). The number of possible states of the tumor is thus . In typical applications one is interested in probability distributions over this state space that are far from stationary. Extrapolating the future course of a given tumor requires the computation of transient distributions. However, since the age of a tumor and thus the time point of an observation of the Markov chain is generally unknown, we study the transient distribution integrated over an exponentially distributed observation time, called the time-marginal distribution. It can be obtained as the solution of a large linear system. Today at least genes are known to drive tumor progression, cf. (Bailey et al. 2018), i.e., a single distribution in would require the storage of more entries than there are atoms in the observable universe. This phenomenon is called state-space explosion (Buchholz and Dayar 2007) and renders classical methods for calculating or storing distributions infeasible. This problem is also known in other application areas, e.g., chemical-reaction networks (Anderson et al. 2010), the chemical master equation (Kazeev et al. 2014), Hamiltonian dynamics (Haegeman et al. 2016), queuing networks (Chan 1987), evolutionary dynamics (Niederbrucker and Gansterer 2011) or stochastic neural networks (Yamanaka et al. 1997).
Our main goal is to develop a method for calculating or approximating the marginal distribution. Due to the state-space explosion, methods for calculating the entire marginal distribution in the context of tumor progression have been limited to about genomic events (Schill et al. 2019). In (Gotovos et al. 2021) it is demonstrated that learning models of this size from data is an underspecified problem and that increasing the number d can make inference more robust. This requires a tractable approximation of the marginal distribution. To allow for a probabilistic interpretation of this approximation, the latter should be a probability distribution itself, i.e., all entries should be non-negative, and the sum of all entries should be equal to one. Here, we neglect the non-negativity to ensure an error-controlled approach, see, e.g., (Kim et al. 2013) for an overview. So the following questions need to be answered:
How can we overcome the state-space explosion?
Subsequently, how can we determine marginal distributions?
At the same time, how can we ensure that a solution sums to one?
In order to address question 1 we use so-called low-rank tensor formats. These are known to reduce the cost from exponential to linear in the number of events d provided the distribution exhibits a low-rank structure.
To do so, we model Markov chains of interacting processes as Stochastic Automata Networks (Plateau and Stewart 2000). These allow for a representation of the infinitesimal generator as a sum of Kronecker products. In the context of low-rank tensors, such a representation is referred to as CANDECOMP/PARAFAC (CP) format (Carroll and Chang 1970; Harshman 1970), and the number of elements in this sum is called CP rank. As there are dependencies between individual processes, the transition rates, i.e., entries of the generator, depend on the current state of the Markov chain. We model the transition rates in a separable way where each factor is defined by the current state in one automaton.
Based on the CP representation of the infinitesimal generator, such representations can also be derived for the operator as well as for the right-hand side of the linear system defining the marginal distribution. We use these low-rank tensor representations to overcome the state-space explosion, cf. question 1, and to compute the marginal distribution, cf. question 2
Strategies which have been used so far to calculate the marginal distribution cannot be generalized to low-rank tensors. Existing methods for solving general systems in low-rank tensor formats can be roughly divided into optimization-based approaches and iterative procedures, see, e.g., (Grasedyck et al. 2013) for an overview. Here we focus on iterative ones and discuss related works in Sect. 3.2. In low-rank formats, the effort for storage and execution of arithmetic operations depends, as the name already suggests, on the rank. Arithmetic operations needed in iterative procedures lead to an increase in the representation rank. One way to counteract this rank increase is the so-called truncation, i.e., the approximation of a tensor by one of lower rank. This allows for reducing storage and computational costs and thereby for efficient iterative procedures in low-rank formats. Beside the CP format there are several other low-rank tensor formats known. Here we focus in particular on the hierarchical Tucker format (Hackbusch and Kühn 2009; Grasedyck 2010). A special feature of these compared to the CP format is the possibility of accurate truncation. In (Hackbusch et al. 2008) it is proven that a convergent iterative method which is supplemented by truncation still converges if the approximation error is small enough.
While guaranteeing convergence, iterative methods supplemented by truncation do not ensure that their results sum to one anymore. In order to address question 3, we derive a novel iterative method based on the Neumann series (Dubois et al. 1979) and the uniformization method (Grassmann 1977) using low-rank tensor formats. We verify that its result sums up to one and prove its convergence. In our numerical experiments, we focus on the concept of Mutual Hazard Networks (Schill et al. 2019) for tumor progression. Our experiments illustrate that the marginal distribution can be approximated by low-rank tensors using our new algorithm.
This work is organized as follows. In Sect. 2 we derive the linear system that defines the time-marginal distribution of a continuous-time Markov chain. Starting from the concept of Mutual Hazard Networks and Stochastic Automata Networks explained there, we define a model class of Markov chains describing interacting processes. For these we will then derive low-rank tensor representations for the operator and the right-hand side of the linear system. In Sect. 3 we introduce the concept of tensors. We review the hierarchical Tucker format and discuss related work. Using these formats we derive an iterative method and prove its convergence. In Sect. 4 we perform numerical experiments based on the concept of Mutual Hazard Networks. In Sect. 5 we conclude.
Statement of the problem and modeling
Time-marginal distribution
A continuous-time Markov chain is defined by its state space , its infinitesimal generator and an initial distribution . In this paper we assume that the state space is discrete. The generator is an operator , i.e., a linear mapping from to , which stores the rates of transition from state x to another state y in and the rates of staying in a state x in . By construction each column of sums up to 0.1 The probability distribution as function of the time is defined as the solution of the initial value problem
1 |
i.e., for all . We make the common assumption that every trajectory starts at the same state, i.e., the initial distribution is a canonical unit vector.
Here, we assume that the observation time t is unavailable, such as, e.g., in tumor progression modeling, and thus must be treated as a random variable. Therefore, we are interested in a so-called time-marginal distribution which is independent of the time t and which we will call marginal distribution for brevity. Each entry of the marginal distribution indicates the probability of observing a state at a random time. We follow the common assumption that the sampling time is an exponentially distributed random variable with rate 1, i.e., . Similar approaches can be found, e.g., in Hjelm et al. (2006); Beerenwinkel and Sullivant (2009); Schill et al. (2019); Gotovos et al. (2021). Please note that for general with the same results can be obtained by simply rescaling the time, i.e., . Here, we have
2 |
The following lemma shows that this integral exists.
Lemma 1
Let be the infinitesimal generator of a continuous-time Markov chain over the discrete state space . Then the integral (2) exists and
3 |
Proof
For the existence of the improper integral we show that the spectrum of the operator fulfills . Applying the Gershgorin circle theorem (Gershgorin 1931) for with , we obtain
and thus, . Then the statement follows from direct calculation.
This result can also be obtained by understanding the process as an absorbing Markov chain where from each state a transition with rate 1 enters the absorbing state. Then the time-marginal distribution equals the distribution just before absorption, see also (Gotovos et al. 2021, Proposition 2). From this point of view, the statement follows from the theory of absorbing Markov chains.
Hence, the marginal distribution is defined as the unique solution of a linear system, since the operator is regular.
Next, we specify the class of Markov chains for which we compute the time-marginal distribution . We start with an introduction into a specific type of model used in tumor progression and then generalize this type based on the concept of Stochastic Automata Networks (Plateau and Stewart 2000). All these Markov chains will offer a sparse representation of the infinitesimal generator in order to overcome the state-space explosion.
Modeling tumor progression via Mutual Hazard Networks
As mentioned in the introduction, tumor progression can be modeled as continuous-time Markov chain over a discrete state space . Typically, one is interested in a transient distribution, but the time point of observation, i.e., the age of the tumor, is unknown. Nevertheless, in order to be able to make statements about the probability distribution for all possible tumors, the time-marginal distribution is needed.
For tumor progression, the state space can be modeled as follows. By consideration of d genomic events, as point mutations, copy number alterations, or changes in DNA methylation, each state represents the genotype of a tumor by indicating whether a genomic event has occurred or not. Modelling a state as a vector of length d with , if event i has occurred, and otherwise , the set of all possible states (or tumors respectively) can be represented as
4 |
Thus, the number of possible tumors increases exponentially in the number d of genomic events considered, also known as state-space explosion.
In tumor progression modeling, the transition rates are usually unknown, and therefore certain assumptions have to be made. Here we focus on the concept of Mutual Hazard Networks (Schill et al. 2019) which offers a sparse representation of the infinitesimal generator . In (Schill et al. 2019) the following assumptions define a Mutual Hazard Network:
-
(i)
All events are assumed to occur one after another.
-
(ii)
All events are irreversible, i.e., there are no transitions from state with to with for an event i.
-
(iii)
The occurrence of an event depends on the genotype of the tumor in a separable way. This means that there are mutual effects between events on their rate of occurrence which can be factorized, and each factor is described by a certain event.
-
(iv)
Genomic events that have not occurred yet have no effect on the transition rates of all others.
According to assumption (iii) a first-order Cox proportional hazard model (Cox 1972) is used to specify the transition rates, i.e., there are parameters defining the mutual effects on transition rates from state x to y,
5 |
where x and y only differ in one event i with . To be precise, the parameter is the (multiplicative) effect of state for event j on the transition from to for event i for events . The parameter is then the baseline rate of transition from to for event i. Following (iv) all effects with and are equal to 1, i.e., they are neutral multiplicative effects in (5).
This modeling allows for describing different mutual effects between genomic events on their transition rates, see (iii). Figure 1 (Schill et al. 2019, supplementary material Figure 1) shows the Mutual Hazard Network inferred for breast cancer data.
Each box represents a genomic event relevant to breast cancer and each line a direct effect between them. A dashed line shows a unilateral effect and a solid line identifies a reciprocal effect. In addition, we can distinguish effects based on their nature and magnitude. In the present case, an amplification on the p-arm of the 16th chromosome (event +16p) inhibits a deletion on the p-arm of the 8th chromosome (event -8p) with . Such an inhibitory effect of event j on i is indicated in the network by a red connection (lines without marks) and corresponds to a parameter . Similarly, event +16p promotes an amplification on the q-arm of the 20th chromosome (event +20q) with . A promoting effect of event j on i is indicated by blue connection (lines with circles) in the network and corresponds to a parameter . Neutral effects are represented in Fig. 1 by the absence of edges, e.g., event +16p is not connected to event +1q which means that event +16p does not affect event +1q directly. At the level of parameters this corresponds to a multiplicative effect of . As can already be seen in Fig. 1, in typical applications most of the direct effects between events are assumed to be neutral. Beyond these direct effects, events can also influence each other indirectly. In the example shown, event +16p is not associated with event +1p, so it has a neutral direct effect on event +1p. However, event +16p favors the occurrence of event +20q, which itself inhibits event +1p. Excluding all other effects, event +16p could therefore indirectly have an inhibitory effect on event +1p (despite ).
The diagonal entries of follow directly from (5) since each column sums up to 0, i.e., . Together with the separation in (5) this allows for the following representation of the infinitesimal generator, cf. (Schill et al. 2019):
6 |
Instead of entries, one only has to store entries. In the following, we will observe that the structure in (6) of the operator will allow us to use tensor formats for the calculation of the time-marginal distribution.
While this is a specific model for tumor progression on genomic data, the question arises whether a similar structure of can be used for more general models with a larger range of applications. In general, the progression of disease is not an irreversible process, and symptoms and traits can arise and then abate, such as inflammation (Zhao et al. 2021). They also do not necessarily have to be modeled in a binary way, but can have different levels of severity, such as fever (Johnston et al. 2019). The same formalism can be used to model the attrition and maintenance of technical systems of interacting components (Amoia et al. 1981). To do this, we introduce the concept of Stochastic Automata Networks (Plateau and Stewart 2000) which are known to offer a similar representation for the infinitesimal generator . We note that the Markov chains defined by a Mutual Hazard Network also belong to this class of Stochastic Automata Networks.
Stochastic automata networks
For a continuous-time Markov chain of interacting processes the discrete state space factorizes in a natural way into the state spaces of the individual processes. Each process is itself a Markov chain over its own state space and is called a stochastic automaton . The set of d stochastic automata is called a Stochastic Automata Network, cf. (Plateau and Stewart 2000). The full state space consists of all possible combinations of states in each automaton, i.e., for and it is given by
7 |
Each state specifies its state in each automaton . Already at this point we note overlaps with the concept of Mutual Hazard Networks. Each genomic event i of the Mutual Hazard Network represents a stochastic automaton with local state space . This results in the global state space for d genomic events in (4).
We now consider transitions between states in the full state space , which are given by one or more transitions between states within the individual state spaces .2 It is known that the infinitesimal generator of a Stochastic Automata Network can be represented as a sum of Kronecker products, i.e., for d automata and m transitions it is given by
8 |
cf. (Plateau and Stewart 2000). Again, instead of entries, one only has to store entries.
A specific way to define the matrices is given in (6) for a Mutual Hazard Network, where the number of possible transitions m is equal to the number of events d. Note that the reduction of to d terms in the sum results from the separability of the transition rates in (5).
Based on the formalism of Stochastic Automata Networks we extend the concept of Mutual Hazard Networks. On the one hand, the separability and parameterization should be preserved. On the other hand, an extension of the models and an enlargement of the possible application areas should be made possible by relaxing the assumptions.
Generalized modeling
We focus on a continuous-time Markov chain represented as a Stochastic Automata Network with d automata . As in Sect. 2.3 each automaton has a state space with , and the state space of the network is given by . According to (i) and (iii) we make the following assumptions:
-
(I)
There is only one transition in one automaton at a time.
-
(II)
Each transition depends on the current global state in a separable way, i.e., each transition rate can be factorized, and each factor is described by the current local state in one automaton.
For each automaton we denote the set of transitions in by . Again, we model the mutual effects on local transitions using parameters . Each parameter is the effect of state in automaton on the transition in automaton for and the baseline rate for . Thus, the transition rate from state x to y in , where x and y differ only in automaton and satisfy , can be represented with parameters by
The diagonal entries of the generator are again given by , and thus we specify a data-sparse representation by
9 |
For , is a diagonal matrix containing the effects of all states in automaton on the transition in , i.e., its diagonal entries are given by the parameters for all states in automaton . For , the matrix contains only two non-zero entries: The entry corresponding to is the baseline rate , and the diagonal entry at is . Similar to the representation of for the Mutual Hazard Networks in (6), the number of terms in the sum is given by the overall number of possible transitions in the Markov chain, i.e., . Compared to the general representation (8) for Stochastic Automata Networks, the reduction of terms is possible due to the separability of transition rates (II).
In contrast to the model based on Mutual Hazard Networks, this generalized version allows for different and larger local state spaces (instead of binary for all i) and does not require specific restrictions as irreversibility or specific neutral effects of states. In particular, the automata may be cyclic. While the data-sparse structure of the infinitesimal generator based on certain parameters is preserved. In the numerical experiments, we will then refer to the case of Mutual Hazard Networks, which have already been used in practical simulations for tumor progression (Schill et al. 2019; Gotovos et al. 2021).
Structure of the linear system
As mentioned in Sect. 1, the solution of the linear system
10 |
in Schill et al. (2019) based on classical methods was limited to about automata. In order to overcome the state-space explosion, we need to go beyond classical methods. We have already given a data-sparse representation of , also in the more general case of Sect. 2.4. Note that the identity and the initial distribution with initial state can be written as Kronecker products, where is the identity operator on and is the -th canonical unit vector. Hence, the operator and the right-hand side of (10) have a representation as a short sum of d Kronecker products, which allows for efficient storage. We still need a low-rank method to solve the linear system, which is the subject of the next section.
Low-rank method to compute the marginal distribution
We now compute the marginal distribution in a data-sparse way. To avoid losing this sparsity when performing arithmetic operations we regard our operators and distributions as tensors by interpreting the Kronecker products as tensor products.
Low-rank tensor formats
We view tensors as multidimensional generalizations of vectors and matrices, i.e., of one-dimensional and two-dimensional tensors.
Definition 1
(tensor) Let and be a Cartesian product of discrete index sets . An object is called a tensor of dimension d. Each direction is called a mode of , and the cardinality of the i-th index set is called the i-mode size.
In our case, the index set corresponds to the state space , our distributions , are tensors of dimension d, and the automata correspond to the modes with sizes .
In the language of tensors, the structure of in (9) is an example of the so-called CANDECOMP/PARAFAC (CP) format introduced in Carroll and Chang (1970); Harshman (1970).
Definition 2
(CP format) A tensor has a CP representation if there exist such that
11 |
Then is called the CP representation rank, and the are called the CP factors of . The minimal such that has such a CP representation (11) is called the CP rank of .
The infinitesimal generator in (9) has CP representation rank . The identity as well as the right-hand side have CP rank 1. A core advantage of the CP format is the data sparsity in case of small representation rank r: The representation (11) of a tensor has storage complexity in in contrast to for the full tensor. We use the -notation for storage and computational costs to describe that costs in grow asymptotically no faster than for a constant and a function f depending on certain value .
Since we want to compute the marginal distribution , we have to solve a linear system whose operator and right-hand side have a CP representation each. For an operator with CP rank it is unknown how to calculate its inverse analytically. Hence, we need a solver and arithmetic operations to compute the solution numerically. Performing arithmetic operations such as adding two tensors and applying an operator results in an increase in the representation rank. In the CP format, this increase in representation ranks can be traced easily: For example, if we add two tensors and with representation ranks and by appending the CP factors of to those of , the sum already has representation rank . Similarly, applying a CP operator with representation rank to a CP tensor with representation rank , by applying the operator factor by factor to the CP tensor, results in a CP tensor with representation rank .
To counteract this increase in the representation rank, arithmetics in low-rank formats is supplemented by a so-called truncation, i.e., approximating a tensor with one of lower representation rank. As the set of tensors with CP rank at least r is not closed for , low-rank approximation within the CP format is an ill-posed problem, see de Silva and Lim (2008), and usually optimization-based methods are used for this purpose. However, we overcome this drawback by using tensor formats that allow for truncation based on the singular value decomposition of matrices.
To do so, a high-dimensional tensor is reshaped into a matrix by selecting modes defining its rows, while all others define its columns. The resulting matrix, called matricization, is defined, following (Grasedyck 2010), as follows.
Definition 3
(matricization) Let and with and . The matricization of corresponding to t is defined as with and
for all . In particular .
The matricizations of a three-dimensional tensor corresponding to the single modes and are visualized in Fig. 2.
For a matricization of a tensor, the classical singular value decomposition can be used for low-rank approximation. So-called tree tensor formats make use of this observation. A popular one is the tensor train format which was first introduced to the numerical analysis community in Oseledets and Tyrtyshnikov (2009). It is also known in other areas as matrix product states (White 1992; Östlund and Rommer 1995) or as linear tensor network (van Loan 2008).
Here we focus on the hierarchical Tucker format which was first introduced in Hackbusch and Kühn (2009) and further analyzed in Grasedyck (2010). As the name suggests, the hierarchical Tucker format is based on a hierarchical subdivision of the modes . This bisection of the modes is described by a binary dimension tree. The set of all modes is distributed from the root to the children until only single element subsets are left in the leaves. Formally, a dimension tree can be defined as follows:
Definition 4
(dimension tree) A dimension tree for dimension is a binary tree with nodes labeled by non-empty subsets of . Its root is labeled with D and each node t (identified with its label) satisfies one and only one of the following conditions:
is a leaf of and labeled by a single element subset .
is an inner node of and has exactly two children which fulfill and .
An example for a dimension tree with dimension is shown in Fig. 3.
The rank of a representation in the hierarchical Tucker format depends on the bisection, i.e., on the dimension tree. For each node t there is a rank component corresponding to the matrix rank of the matricization . Thus, the representation rank in the hierarchical Tucker format is a tuple depending on the tree and will be denoted by . Again, the minimal representation rank is called the hierarchical Tucker rank.
Given a dimension tree , truncation can be performed in a quasi-optimal way, cf. (Grasedyck 2010). There it is shown that the error in the -norm caused by truncation of a tensor to rank is bounded by
12 |
where is the m-th singular value of the matricization , is a best rank- approximation, and is a small constant. This error bound allows for truncation with guaranteed accuracy. Thus, truncating after arithmetic operations allows us to perform iterative methods in an efficient way and preserves their convergence (Hackbusch et al. 2008).
Inspired by (Hackbusch 2012, Chapter 13), Table 1 lists some of these operations, as well as their respective cost.
Table 1.
In the hierarchical Tucker format, the storage complexity grows only linearly in the dimension d. This allows for efficient storage and overcomes the state-space explosion provided we have low ranks. In our case the dimension of the distribution tensors is equal to the number d of automata, and the mode sizes are the numbers of states for each automaton .
It is possible to convert a CP representation of rank r easily to a hierarchical Tucker one, where all rank components are bounded by r independent of the dimension tree, cf. (Hackbusch 2012). In general, however, the rank in the hierarchical Tucker format depends on the selected bisection, i.e., on the dimension tree. Thus, the question arises how to choose the dimension tree in order to obtain a low rank. For further explanation on the choice of tree we refer the reader to more advanced papers (Grasedyck and Hackbusch 2011; Ballani and Grasedyck 2014). We will discuss this issue for the marginal distribution in Sects. 4.2 and 4.3 using some numerical experiments.
Related work
For large Markov chains, the data-sparse structure of the operator has already been exploited for fast matrix-vector applications (Buchholz and Dayar 2007). However, this allowed only the operator, but not the distributions, to be stored in a data-sparse form. The distributions were stored entry by entry and thus suffer from the state-space explosion. In Schill et al. (2019) a strategy based on a splitting of the operator in a diagonal matrix and a strictly lower triangular matrix was presented. Using and the Neumann series, the marginal distribution is given by
13 |
Note that this equation requires inverting the diagonal operator or solving the corresponding linear system. In contrast to the matrix case, inverting a diagonal operator in low-rank tensor formats is not straightforward, and it is unclear whether the solution itself has a low-rank structure. Solving each linear system involving would result in solving systems with a comparable complexity as the original one. Thus, the strategy in (13) cannot easily be transferred to low-rank tensors to overcome the state-space explosion.
Alternatively, Gotovos et al. (2021) recently proposed a strategy for learning the parameters of a Mutual Hazard Network without computing the marginal distribution for all states. Given a data set of observations, optimizing the log-likelihood requires the marginal probability of each state observed, which is represented as sum of probabilities for all possible transition sequences leading to this state. For an observation with m mutations present there are m! possible sequences. In order to deal with the sheer number of sequences, a stochastic approximation based on a Metropolis-Hastings algorithm is used. Instead of restricting the state space, here we want to exploit the given low-rank tensor structures of the model, cf. Sect. 2.5, to overcome the state-space explosion.
We briefly review existing strategies to compute distributions using low-rank tensor formats. The CP format has been used extensively to approximate, in particular, stationary distributions, e.g., in Kulkarni (2011), and to derive conditions for their existence, e.g., in Fourneau (2008). In Benson et al. (2017) stationary distributions for random walks are computed via an eigenvalue problem using the CP format. The tensor-train format was successfully used for the computation of, e.g., transient distributions (Johnson et al. 2010), mean times to failure (Robol and Masetti 2019), and stationary distributions (Bolten et al. 2018; Kressner and Macedo 2014). There mainly optimization-based approaches were presented. In Buchholz et al. (2016) the hierarchical Tucker format was applied to reduce the storage cost for distributions and the computational cost for performing basic operations. Continuing in Buchholz et al. (2017) adaptive truncation strategies for the computation of stationary distributions using iterative methods were presented. In Kressner and Macedo (2014) a power iteration based on a formulation of the stationary solution by an eigenvalue problem was derived, where after each application of the operator in the power iteration the current approximation was rescaled to ensure that it sums up to one.
However, the time-marginal distribution we are interested in has not yet been studied in the context of low-rank tensors. For this reason, we present such a method in the following section.
Low-rank method
We now make use of low-rank tensor formats to approximate the marginal distribution as the solution of (10). Since the exact solution is a probability distribution, its entries sum up to one, i.e., it fulfills
14 |
where is the tensor of all ones. To allow for a probabilistic interpretation of an approximation of , we need to treat (14) as an additional constraint. To this end, we now present an iterative method based on the Neumann series (Dubois et al. 1979) and the uniformization method (Grassmann 1977). The following lemma holds true for any discrete-state continuous-time Markov chain.
Lemma 2
Let be any infinitesimal generator over any discrete state space and any initial distribution. For the marginal distribution can be represented as a Neumann series,
15 |
and the series converges.
Proof
Due to Lemma 1 the marginal distribution can be represented as
In order to use the Neumann series for inverting , it is sufficient to show that the spectral radius . Applying the Gershgorin circle theorem (Gershgorin 1931) leads to
Thus, the spectral radius is bounded by
For and we obtain using :
Since the spectral radius is smaller than 1, the Neumann series converges with
Alternatively we can derive (15) using the uniformization method. Here, the idea is to describe a continuous-time Markov chain by a discrete-time Markov chain with a time increment that is exponentially distributed. With this interpretation, we write the time-dependent probability distribution in (1) as
for a time . Note that is the transition probability matrix of a discrete-time Markov chain, since is a bound on the diagonal entries of . Marginalization of time similar to (2) and substitution leads to (15):
where denotes the gamma function.
We now discuss approximation strategies for (15). A natural approximation would be
16 |
for which is an approximation from below (entry by entry), i.e., for all and all . Based on the properties of the Neumann series this sequence converges linearly to , but satisfies the normalization condition (14) only approximately for . As is a transition probability matrix, its application to a probability distribution leads to a probability distribution again satisfying (14), and thus
17 |
for all . By scaling each element of the sequence, we obtain
18 |
for , which is now an approximation to that satisfies (14). We prove its linear convergence to in the following theorem for any discrete-state continuous-time Markov chain.
Theorem 1
Let be any infinitesimal generator over any discrete state space and any initial distribution. Let further be the solution of the corresponding linear system (10), and be defined by (18) for all . Then converges linearly to as k approaches infinity, i.e.,
where .
Proof
For , and any we have
Furthermore we obtain
Since , the linear convergence follows.
So far, we have not needed any additional assumptions on the Markov chain. However, we are particularly interested in high-dimensional problems and want to overcome the state-space explosion using low-rank tensor methods. Therefore, we now assume that both, the generator and the initial distribution , have a low-rank tensor representation, as derived in Sect. 2. In order to employ low-rank tensor formats, we supplement the iterative method corresponding to (18) by truncation. As shown in Hackbusch et al. (2008), a convergent iterative method combined with truncation still converges if the truncation error is sufficiently small. However, the truncation could change the normalization of the iteration. Therefore, we replace the scaling with by dividing by as shown in Algorithm 1. The procedure should be continued until the norm of the relative residual is smaller than a given tolerance .
Note that when using the unnormalized approximation by replacing the normalization by with , the norm of the relative residual turns out not to be a useful measure and one must also consider the distance of and . By taking the relative residual of the normalized version, this is done implicitly.
Our sequence is defined as a normalized version of , which is an approximation to from below (entry-by-entry). By estimating the Poisson distributed error of , one can obtain
for as unnormalized approximation to from above (entry-by-entry) which also converges linearly to . In numerical experiments (not further discussed here), and also its normalized version turned out to be unpromising, since significantly more iteration steps were required to achieve the same approximation quality measured by the relative residual.
In general, Algorithm (I) does not require assumptions or (II) but and to have appropriate low-rank tensor representations and an upper bound . In general, computing all diagonal entries of is in . Nevertheless, here we focus on Markov chains satisfying (I) and (II) with an infinitesimal generator in (9) depending on parameters . This allows for the following inexpensive upper bound:
19 |
which can be computed in ,3 where is the maximum number of possible transitions in an automaton. In the case of strongly varying parameters , (19) may be a significant overestimation for the diagonal entries of . The question of how to determine a tighter bound with polynomial effort using low-rank tensor formats will be dealt with in future work.
Numerical experiments
We illustrate our method for the computation of time-marginal distributions in numerical experiments based on the model of Mutual Hazard Networks with d automata (genomic events). In this case, the parameters can be summarized in a matrix .
Setting
Construction of synthetic parameters
We consider d automata and parameters with a particular block-diagonal form. Each quadratic block of size characterizes a subset of b automata which directly affect one another. We draw all within each block so that their logarithmic values are normally distributed with mean and standard deviation . Unless stated otherwise, all other parameters are exactly 1 (logarithmic values of 0), which corresponds to a neutral direct effect between automata of different blocks. Based on these blocks we study three types of parameters:
(Strict) block structure: There are direct effects, i.e., parameters , only between automata in the same block.
Neighbor block structure: In addition to the effects within each block in (B1), there are direct effects between randomly chosen automata in neighboring blocks. The parameters are drawn such that their logarithmic values are normal distributed with mean 0 and and the automata (i, j) are uniformly randomized in neighboring blocks. For each pair of neighboring blocks 4 random effects.
Neighbor block structure with additional random effects: Besides the effects in (B2), there are direct effects between randomly chosen automata of different blocks. Again the corresponding parameters are drawn such that their logarithmic values are normal distributed with mean 0 and . The choice of (i, j) is also uniformly randomized. For each parameter matrix we add 8 random effects.
The blocks in (B1) correspond to non-interacting subsystems of the process. In biological models, these often represent well-defined, distinct pathways of genes that regulate specific functions of the cell such as the cell cycle, apoptosis and growth (Sanchez-Vega et al. 2018). Note that even if subsystems do not interact, it is not possible to trivially express the solution as a Kronecker product of solutions per block, since correlations are induced by the observation time acting as a confounding variable, see also (Gotovos et al. 2021, Section 3.2). Finding such blocks or groupings, e.g., during parameter determination, can be important for understanding and treating tumors. However, in practice, the behavior of a biological system can rarely be separated into well-defined subsystems that do not interact, so (B1) is an oversimplification. Hence, we add interactions between different subsystems in (B2) and (B3). For parameters of type (B2) all automata or events already interact indirectly. For example, if event favors the occurrence of event , i.e., , and at the same time event inhibits the occurrence of event , i.e., , then event also indirectly inhibits event even if the direct effect of on is neutral. We reduce the standard deviation when sampling the effects between different blocks to simulate that the influences between different blocks are much smaller than those within.
Default settings for the algorithm
For the application of Algorithm 1 we always choose the bound as in (19), which for Mutual Hazard Networks can be reduced to
20 |
Unless stated otherwise, we use a canonical balanced dimension tree for the application of the hierarchical Tucker format, i.e., the automata are assigned to the leaves following their ordering, see Fig. 4. We perform all experiments for 100 randomly generated sample parameters for each combination parameter type, number of blocks and number d of automata. We compute low-rank approximations of the marginal distribution using Algorithm 1 with a maximum relative truncation error , see (12). The algorithm stops when the norm of the relative residual is smaller than a tolerance value of . The mean values we compute are arithmetic means. In the following, we call the representation rank of this approximation tensor an approximation rank of .
Study of singular values
One fundamental assumption for solving linear systems using low-rank tensor methods is that a solution can be well approximated by a tensor of low rank. According to the error bound in (12), the truncation error is determined by the singular values of the matricization corresponding to the chosen dimension tree. A fast decay of those singular values indicates that a tensor can be approximated accurately by one of low rank.
To analyze this issue, we solve (10) using classical matrix methods of MATLAB (Natick 2019) and compute the singular values of the corresponding matricization using the htucker toolbox (Kressner and Tobler 2014). Figure 4 shows the decay of the singular values of each matricization corresponding to the canonical balanced tree for automata and parameters of type (B1) with 2 blocks of size each. For each node the semi-logarithmic plot displays the means of the singular values over 100 samples.
We observe that the singular values exhibit an exponential decay. The two matricizations closest to the root are transposes of each other, and therefore their singular values are identical. The smallest 5 of the 16 singular values are indistinguishable from zero and therefore cannot be displayed in the semi-logarithmic plot. For other standard deviation , we observed a similar drop in singular values (not shown here). The exponential decay of the singular values indicates that the marginal distribution can be well approximated with low rank.
Study of the tree structure
In general, the rank in the hierarchical Tucker format depends on the structure of the tree. In the following, we study how the ordering of the automata in the leaves of the tree affects the low-rank approximability. We preserve the balanced binary structure of the tree because this is advantageous for parallelization, cf. (Grasedyck and Löbbert 2018).
We already studied the singular values of the matricizations corresponding to the canonical balanced tree, see Fig. 4. We now change only the arrangement of the automata in the leaves of the tree and compute the marginal distribution again. The decay of the singular values for each matricization is shown in Fig. 5, where the semi-logarithmic plot at each vertex displays the means of the singular values over 100 samples.
Comparing Figs. 4 and 5, we observe that the choice of the canonical dimension tree, i.e., the original ordering in the leaves, results in a significantly faster decay of the singular values close to the root. Figure 5 shows that the singular values closest to the root also have an exponential decrease, but at a much slower rate. Note that 2 blocks of size imply that the automata interact directly with one another. The same holds for the automata . There are no direct interactions between automata of different blocks. Hence, the modified balanced tree separates strongly interacting automata, which explains the slower decay of the singular values in level 1 (level 0 being the root). In contrast, the canonical balanced tree separates weakly interacting automata, leading to a faster decay of the singular values. We will make similar observations when studying the approximation rank. Based on this observation, the question arises whether and how a suitable dimension tree can be determined. Ballani and Grasedyck (2014) developed a black box method to determine a dimension tree for general low-rank problems. In our concrete model problem, however, the parameters already seem to give a-priori information about suitable structures. How exactly a dimension tree can be constructed based on the parameters is, however, not a trivial question and requires a more detailed investigation, which would exceed the scope of this paper. In the following, we will focus on the investigation of the presented method.
Comparison with a matrix-based method
We check our method in Algorithm 1 against a classical method based on matrix operations presented in Schill et al. (2019), see (15). To do so, we use Mutual Hazard Networks computed there and the corresponding infinitesimal generators . The parameters are optimized based on data for breast cancer, colorectal cancer, renal cell carcinoma and glioblastoma which represent mutual effects between and 20 genomic events each. The specific parameters and their interpretation can be found in Schill et al. (2019) and its supplementary material. The network for breast cancer is also shown in Fig. 1. We use the settings in Sect. 4.1.2 for Algorithm 1.
We compare the result of Algorithm 1 using tensor methods and of (13) based on matrix operations with respect to the relative residual and the distance of both results. We measure the distance using the relative Euclidean distance and the Kullback Leibler (KL) divergence (Kullback 1997) from to . The KL divergence from to is defined as
21 |
and measures how differs from with respect to . Since is only an approximation of the marginal distribution , there might be some negative values which are absolutely small. To deal with this problem, we use the convention and cut off to zero. Besides, we list the costs for storing and , respectively, in values which have to be stored for each solution. Table 2 summarizes the values for four different Mutual Hazard Networks corresponding to different types of tumors.
Table 2.
Disease | d | Relative residual | Relative residual | Relative Euclidean distance | Absolute KL divergence | Storage cost | Storage cost |
---|---|---|---|---|---|---|---|
Breast cancer | 10 | 1.7e17 | 8.8e05 | 3.3e04 | 3.3e07 | 1024 | 3052 |
Colorectal cancer | 11 | 2.9e17 | 9.8e05 | 3.5e04 | 3.1e07 | 2048 | 4396 |
Renal cell carcinoma | 12 | 1.1e17 | 9.8e05 | 1.2e04 | 1.4e06 | 4096 | 4553 |
Glioblastoma | 20 | 1.6e17 | 9.7e05 | 9.7e05 | 4.4e06 | 1,048,576 | 44,065 |
By construction, the residuals for of Algorithm 1 are slightly below the requested tolerance . For smaller values of (not shown in Table 2) smaller relative residuals for can be reached by increasing the number of iteration steps. The results of (13) solve the equation nearly exactly. The relative Euclidean distances as well as the KL divergences from to have very small absolute values. This implies that both results describe nearly the same distribution. Calculating the KL divergences for smaller cut-off parameters , we obtain the same values for the first three networks as with . For the last (and largest) network, if , we observe some negative entries with very small absolute values, i.e., . This can be explained by the truncation used in Algorithm 1 to reduce the storage and computational complexity. Hence, we need for this network. Calculating KL divergences requires the explicit evaluation of all entries of the tensors , since an efficient method for the entry-wise application of the logarithm within low-rank tensor formats is unknown. This limits the computation of KL divergences to tensors of small dimension d. Comparing the storage costs for and , we observe an overhead of storing the tensor solution compared to the vector one for diseases with . However, already for the storage cost for the vector solution explodes, whereas the cost for the tensor solution remains moderate with about of the cost for .
In Fig. 6, we further modeled the storage requirements for potential vector and tensor solutions as a function of d up to 50 in a semi-logarithmic plot. For this purpose, we made the simplified assumption of constant ranks for the hierarchical Tucker representation of with .
Again, we note an overhead of storing the tensor compared to the vector representation for , but for the tensor representation is preferable. Figure 6 shows the exponential increase of storage cost for the vector representation in the number of automata, i.e., the state-space explosion. In contrast the cost for the hierarchical Tucker representation grows only linearly in d as expected from Table 1. A low representation rank is essential in order to overcome the state space explosion, as it enters cubically into the storage cost. We will now investigate this in more detail.
Study of the approximation rank
The rank in a hierarchical Tucker format is a tuple depending on the underlying tree . For a simple comparison of tensor representations, we consider the effective rank , which we define such that the storage cost for the representation equals the cost to store one with rank . In doing so, we round up to the nearest integer.
We compute low-rank tensor approximations of the marginal distribution using Algorithm 1 as described in Sect. 4.1. We plot the effective ranks as a function of the number d of automata for fixed maximum relative truncation error and fixed tolerance in Fig. 7a, and as a function of for fixed d in Fig. 7b. Both plots show mean values of effective ranks for 100 samples of parameters each. (In both plots the statistical errors, i.e., the variations between samples, are very small and therefore not shown.)
Figure 7a displays effective ranks for parameters of the three different types. The parameters of type (B1) have 8 and 4 blocks, respectively, and those types (B2) and (B3) have 4 blocks. The size of each block increases in the number d of automata, i.e., for 8 blocks the size of each block is and for 4 blocks it is . The automata belonging to a block are each grouped by the tree structure as in Fig. 4.
We observe that is approximated to a tolerance of with low rank in all cases. For separated blocks of type (B1), drawn in blue, the effective rank increases slightly in the number d of automata. Since the number of blocks is kept constant, also the size of each block increases in d. For this reason, especially for 4 blocks, there is a superposition of the effects of increasing d and increasing block size on the effective rank. This becomes particularly clear in the comparison between 8 and 4 blocks per parameter. In case of 8 blocks, the effective rank increases much more flatly and seems to be almost independent of the number d of automata constrained by . The larger 4 blocks always combine the automata of two smaller blocks, i.e., they are twice as large. We thus hold that although the approximation ranks increase slightly with the number d of automata as well as the number b of directly interacting automata, they remain comparatively small, provided that the partitioning in the dimensional tree is preserved. As in Fig. 5, separating the automata of a block in the dimension tree can lead to an increase in the approximation rank (not shown here). A similar increase can be observed when further effects are added between neighboring blocks, i.e., (B2), or between randomly chosen blocks, i.e., (B3). In Fig. 7a the effective ranks for parameters with additional random effects, drawn in red, both increase in d until 24 and then remain almost constant at around for (B2) or slightly decreases respectively for (B3). This suggests that effects that do not fit the structure of the dimension tree can lead to an increase in the approximation rank. In addition, direct effects between neighboring blocks appear to have less impact on the approximation rank than those between more distant ones. For parameters of type (B3), we observe more significant differences between effective ranks for different samples. For some, the effective ranks are significantly above the mean, for some below it. This supports the conjecture that the location (i, j) of the direct effects affects the approximation rank. However, if only a few of these effects are present, the marginal distribution can still be approximated with comparatively low approximation rank. These results suggest that neither the number d of automata nor the size of the blocks alone are responsible for an increase in rank, but that especially the distribution of the automata in the dimension tree has a large effect. How to construct an appropriate tree in order to keep ranks low using a-priori information on the parameters is a topic of ongoing research.
The semi-logarithmic plot in Fig. 7b displays effective ranks for parameters of type (B1) for automata and 4 blocks of size as a function of the maximum relative truncation error for different tolerances . As the tolerance is lowered, the approximation of corresponds to a particular stage during the iteration. We observe that all effective ranks are nearly constant in the maximum relative truncation error and the tolerance for . For , the effective rank is slightly lower and for all other tolerance values it is about . This observation suggests, on the one hand, that can be accurately approximated with a tensor of effective rank , since a more accurate truncation, i.e., smaller , has only very small impact on the approximation ranks. On the other hand, the fact that the ranks are low and nearly independent of the tolerance value indicates that the ranks during the iteration are also low. This allows not only for efficient storage of the resulting approximation but also for efficient computation using Algorithm 1. For parameters of types (B2) and (B3), we observed that the increase in the approximation rank is further amplified for smaller maximum relative truncation errors . Therefore, a suitable truncation accuracy seems to become more important for parameters with more direct effects between blocks. For Fig. 7a, the maximum relative truncation error is chosen to be comparatively small, but the approximation ranks remain relatively small for all numbers d of automata and types of parameters. During the iteration itself, the effective ranks remained almost constant (after a start phase) similar as shown in Fig. 7b.
Study of the speed of convergence
Having studied the low-rank structure of the distribution numerically, we now consider the convergence speed of our method. In Theorem 1 we proved that the iteration sequence converges linearly to the marginal distribution. If the truncation error is small enough, then the convergence of the method combined with truncation is preserved, cf. (Hackbusch et al. 2008). We will now see that this theoretical result also holds in practice.
To do this, we look at the decay of the relative residual as a function of the iteration steps. Again we use 100 parameter samples of type (B1) with 4 blocks for and 32 automata each. Figure 8a and b show semi-logarithmic plots of the norm of the relative residual as a function of the iteration step. Figure 8a displays the mean value of the relative residual and Fig. 8b additionally the corresponding box plot illustrating the variances for automata.
We observe a linear convergence of the method for all values of d. For larger number d of automata the convergence slows down. A similar behavior can be observed for larger standard deviations of the logarithmic parameters , where a larger means that also the parameters are more widely dispersed around 1. Both observations can be explained by the fact that (and our upper bound in (20)) increases in d and , and thus the bound on the convergence rate is closer to one, cf. Theorem 1. In Fig. 8b the ranges for given by the boxes are small, which indicates that there are only a few outliers given by the whiskers. We have confirmed our results using different numbers d of automata and blocks (not shown). In our tests we observe that the number of iteration steps needed to achieve a certain tolerance grows linearly in the number d of automata but is nearly independent of the block size. This indicates that the number of iteration steps is independent of changes in the ordering of automata and consequently of changes in the rank.
Conclusion and future work
Inspired by current research in tumor progression models, we considered a class of continuous-time Markov chains that describe interacting processes, e.g., tumor progression models. Typically the age of a tumor at its diagnosis is unknown. For this reason, the transient distribution integrated over the exponentially distributed observation time is required. This so-called time-marginal distribution is uniquely defined as the solution a large linear system and suffers from the problem of state-space explosion. We modeled this class of Markov chains with separable transition rates factorizing according to the current state using Stochastic Automata Networks. This modeling enabled us to obtain a low-rank tensor representation of the operator and the right-hand side of this linear system. Based on these low-rank tensor representations, we derived an iterative method to compute a low-rank tensor approximation of the time-marginal distribution and hence were able to overcome the state-space explosion. The method guarantees that the entries of the approximation sum up to one as required for a probability distribution. We proved the convergence of the method. In numerical experiments focused on the concept of Mutual Hazard Networks we illustrated that the time-marginal distribution is well approximated with low rank. The method allows for consistently low ranks during the iteration, and linear convergence was observed independently of the number of processes/automata.
A probability distribution, in addition to being normalized to one, must be non-negative. An approximation of a probability distribution should also satisfy this condition. In numerical experiments we observed that our method preserves this non-negativity for small numbers of automata if the truncation is sufficiently accurate. How to guarantee non-negativity and at the same time convergence for high numbers of automata will be part of our future research. Moreover, we observed that the approximation rank for the time-marginal distribution depends strongly on the structure of the dimension tree and on the effects between automata. To minimize the approximation rank we plan to develop a strategy to determine an optimal dimension tree a-priori.
Acknowledgements
We thank Tim A. Werthmann for his critical reading of and suggestions for this article.
Funding
Open Access funding enabled and organized by Projekt DEAL. This work was partially supported by the German Research Foundation (DFG) grant SFB/TRR-55 “Hadron Physics from Lattice QCD”, SPP-1886 “Polymorphic uncertainty modelling for the numerical design of structures”, TRR-305 “Striking a moving target: From mechanisms of metastatic colonization to novel systemic therapies” and GR-3179/6-1 “Tensorapproximationsmethodem zur Modellierung von Tumorprogression”.
Data availability
For the corresponding data and code, please contact the authors.
Declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Footnotes
Note that in other areas it is also common to restrict to have row sum 0. Here we use column sum 0 consistently with tumor progression literature. All results can be easily transferred by transposing.
For brevity, we do not introduce the concept of functional transitions and restrict ourselves to constant ones, since the separability in (5) can be exploited for reformulation as constant transitions. We refer the interested reader to (Plateau and Stewart 2000) for further details.
Strictly speaking, the computational cost of the method is therefore quadratic, rather than linear, in d. However, only needs to be precomputed once, the computational cost of which is negligible.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- Amoia V, Micheli GD, Santomauro M. Computer-oriented formulation of transition-rate matrices via Kronecker algebra. IEEE Trans Reliab R. 1981;30(2):123–132. doi: 10.1109/tr.1981.5221004. [DOI] [Google Scholar]
- Anderson DF, Craciun G, Kurtz TG (2010) Product-form stationary distributions for deficiency zero chemical reaction networks. Bull Math Biol 72(8):1947–1970. 10.1007/s11538-010-9517-4 [DOI] [PubMed]
- Bailey MH, Tokheim C, Porta-Pardo ELD. Comprehensive characterization of cancer driver genes and mutations. Cell. 2018;173(2):371–385.e18. doi: 10.1016/j.cell.2018.02.060. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ballani J, Grasedyck L. Tree adaptive approximation in the hierarchical tensor format. SIAM J Sci Comput. 2014;36(4):A1415–A1431. doi: 10.1137/130926328. [DOI] [Google Scholar]
- Beerenwinkel N, Sullivant S. Markov models for accumulating mutations. Biometrika. 2009;96(3):645–661. doi: 10.1093/biomet/asp023. [DOI] [Google Scholar]
- Beerenwinkel N, Schwarz RF, Gerstung M, Markowetz F. Cancer evolution: mathematical models and computational inference. Syst Biol. 2014;64(1):e1–e25. doi: 10.1093/sysbio/syu081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Benson AR, Gleich DF, Lim LH. The spacey random walk: a Stochastic process for higher-order data. SIAM Rev. 2017;59(2):321–345. doi: 10.1137/16M1074023. [DOI] [Google Scholar]
- Bolten M, Kahl K, Kressner D, Macedo F, Sokolović S. Multigrid methods combined with low-rank approximation for tensor-structured Markov chains. ETNA Electron Trans Numer Anal. 2018;48:348–361. doi: 10.1553/etna_vol48s348. [DOI] [Google Scholar]
- Buchholz P, Dayar T. On the convergence of a class of multilevel methods for large sparse Markov chains. SIAM J Matrix Anal Appl. 2007;29(3):1025–1049. doi: 10.1137/060651161. [DOI] [Google Scholar]
- Buchholz P, Dayar T, Kriege J, Orhan MC (2016) Compact representation of solution vectors in kronecker-based Markovian analysis. In: Quantitative evaluation of systems. Springer, New York, pp 260–276, 10.1007/978-3-319-43425-4_18
- Buchholz P, Dayar T, Kriege J, Orhan MC. On compact solution vectors in Kronecker-based Markovian analysis. Perform Eval. 2017;115:132–149. doi: 10.1016/j.peva.2017.08.002. [DOI] [Google Scholar]
- Carroll JD, Chang JJ. Analysis of individual differences in multidimensional scaling via an n-way generalization of “Eckart-Young” decomposition. Psychometrika. 1970;35(3):283–319. doi: 10.1007/bf02310791. [DOI] [Google Scholar]
- Chan RH. Iterative methods for overflow queueing models i. Numer Math. 1987;51(2):143–180. doi: 10.1007/bf01396747. [DOI] [Google Scholar]
- Cox DR. Regression models and life-tables. J R Stat Soc Ser B (Methodol) 1972;34(2):187–202. doi: 10.1111/j.2517-6161.1972.tb00899.x. [DOI] [Google Scholar]
- de Silva V, Lim LH. Tensor rank and the Ill-Posedness of the best low-rank approximation problem. SIAM J Matrix Anal Appl. 2008;30(3):1084–1127. doi: 10.1137/06066518x. [DOI] [Google Scholar]
- Dubois PF, Greenbaum A, Rodrigue GH. Approximating the inverse of a matrix for use in iterative algorithms on vector processors. Computing. 1979;22(3):257–268. doi: 10.1007/bf02243566. [DOI] [Google Scholar]
- Fourneau JM (2008) Product form steady-state distribution for stochastic automata networks with domino synchronizations. In: Computer Performance Engineering, Springer, Berlin, pp 110–124, 10.1007/978-3-540-87412-6_9
- Gershgorin SA. Über die Abgrenzung der Eigenwerte einer Matrix. Izvestija Akademii Nauk SSSR, Serija Matematika. 1931;7(3):749–754. [Google Scholar]
- Gotovos A, Burkholz R, Quackenbush J, Jegelka S (2021) Scaling up continuous-time Markov chains helps resolve underspecification. arXiv:2107.02911
- Grasedyck L. Hierarchical singular value decomposition of tensors. SIAM J Matrix Anal Appl. 2010;31(4):2029–2054. doi: 10.1137/090764189. [DOI] [Google Scholar]
- Grasedyck L, Hackbusch W. An introduction to hierarchical (H-)rank and TT-rank of tensors with examples. Comput Methods Appl Math. 2011;11(3):291–304. doi: 10.2478/cmam-2011-0016. [DOI] [Google Scholar]
- Grasedyck L, Löbbert C. Distributed hierarchical SVD in the hierarchical Tucker format. Numer Linear Algebra Appl. 2018 doi: 10.1002/nla.2174. [DOI] [Google Scholar]
- Grasedyck L, Kressner D, Tobler C. A literature survey of low-rank tensor approximation techniques. GAMM-Mitteilungen. 2013;36(1):53–78. doi: 10.1002/gamm.201310004. [DOI] [Google Scholar]
- Grassmann W. Transient solutions in Markovian queueing systems. Comput Oper Res. 1977;4(1):47–53. doi: 10.1016/0305-0548(77)90007-7. [DOI] [Google Scholar]
- Hackbusch W. Tensor spaces and numerical tensor calculus. Berlin: Springer; 2012. [Google Scholar]
- Hackbusch W, Kühn S. A new scheme for the tensor representation. J Fourier Anal Appl. 2009;15(5):706–722. doi: 10.1007/s00041-009-9094-9. [DOI] [Google Scholar]
- Hackbusch W, Khoromskij BN, Tyrtyshnikov EE. Approximate iterations for structured matrices. Numer Math. 2008;109(3):365–383. doi: 10.1007/s00211-008-0143-0. [DOI] [Google Scholar]
- Haegeman J, Lubich C, Oseledets I, Vandereycken B, Verstraete F. Unifying time evolution and optimization with matrix product states. Phys Rev B. 2016;94(16):165116. doi: 10.1103/physrevb.94.165116. [DOI] [Google Scholar]
- Harshman R (1970) Foundations of the parafac procedure: models and conditions for an ’exploratory’ multimodal factor analysis. In: UCLA working papers in phonetics, phonetics, pp 1–84
- Hjelm M, Höglund M, Lagergren J. New probabilistic network models and algorithms for oncogenesis. J Comput Biol. 2006;13(4):853–865. doi: 10.1089/cmb.2006.13.853. [DOI] [PubMed] [Google Scholar]
- Johnson TH, Clark SR, Jaksch D. Dynamical simulations of classical stochastic systems using matrix product states. Phys Rev E. 2010;82(3):036702. doi: 10.1103/physreve.82.036702. [DOI] [PubMed] [Google Scholar]
- Johnston IG, Hoffmann T, Greenbury SF, Cominetti O, Jallow M, Kwiatkowski D, Barahona M, Jones NS, Casals-Pascual C. Precision identification of high-risk phenotypes and progression pathways in severe malaria without requiring longitudinal data. NPJ Digital Med. 2019 doi: 10.1038/s41746-019-0140-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kazeev V, Khammash M, Nip M, Schwab C. Direct solution of the chemical master equation using quantized tensor trains. PLoS Comput Biol. 2014;10(3):e1003359. doi: 10.1371/journal.pcbi.1003359. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kim J, He Y, Park H. Algorithms for nonnegative matrix and tensor factorizations: a unified view based on block coordinate descent framework. J Global Optim. 2013;58(2):285–319. doi: 10.1007/s10898-013-0035-4. [DOI] [Google Scholar]
- Kressner D, Macedo F (2014) Low-rank tensor methods for communicating Markov processes. In: Quantitative evaluation of systems. Springer, New York, pp 25–40, 10.1007/978-3-319-10696-0_4
- Kressner D, Tobler C. Algorithm 941. ACM Trans Math Softw. 2014;40(3):1–22. doi: 10.1145/2538688. [DOI] [Google Scholar]
- Kulkarni VG. Introduction to modeling and analysis of Stochastic systems. New York: Springer; 2011. [Google Scholar]
- Kullback S. Information theory and statistics. Mineola: Dover Publications; 1997. [Google Scholar]
- Natick M (2019) MATLAB. https://de.mathworks.com/products/matlab.html, version 9.6 (R2019a)
- Niederbrucker G, Gansterer WN (2011) A fast solver for modeling the evolution of virus populations. In: Proceedings of 2011 international conference for high performance computing, networking, storage and analysis on - SC ’11, ACM Press, New York, 10.1145/2063384.2063483
- Östlund S, Rommer S. Thermodynamic limit of density matrix renormalization. Phys Rev Lett. 1995;75(19):3537–3540. doi: 10.1103/physrevlett.75.3537. [DOI] [PubMed] [Google Scholar]
- Oseledets IV, Tyrtyshnikov EE. Breaking the curse of dimensionality, or how to use SVD in many dimensions. SIAM J Sci Comput. 2009;31(5):3744–3759. doi: 10.1137/090748330. [DOI] [Google Scholar]
- Plateau B, Stewart WJ (2000) Stochastic automata networks. In: International series in operations research and management science, Springer, New York, pp 113–151, 10.1007/978-1-4757-4828-4_5
- Robol L, Masetti G (2019) Tensor methods for the computation of MTTA in large systems of loosely interconnected components. https://arxiv.org/abs/1907.02449
- Sanchez-Vega F, Mina MJA. Oncogenic signaling pathways in the cancer genome atlas. Cell. 2018;173(2):321–337.e10. doi: 10.1016/j.cell.2018.03.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schill R, Solbrig S, Wettig T, Spang R. Modelling cancer progression using Mutual Hazard Networks. Bioinformatics. 2019;36(1):241–249. doi: 10.1093/bioinformatics/btz513. [DOI] [PMC free article] [PubMed] [Google Scholar]
- van Loan CF (2008) Tensor network computations in quantum chemistry. www.cs.cornell.edu/cv/OtherPdf/ZeuthenCVL.pdf
- White SR. Density matrix formulation for quantum renormalization groups. Phys Rev Lett. 1992;69(19):2863–2866. doi: 10.1103/physrevlett.69.2863. [DOI] [PubMed] [Google Scholar]
- Yamanaka K, Agu M, Miyajima T. A continuous-time asynchronous Boltzmann machine. Neural Netw. 1997;10(6):1103–1107. doi: 10.1016/s0893-6080(97)00006-3. [DOI] [PubMed] [Google Scholar]
- Zhao H, Wu L, Yan G, Chen Y, Zhou M, Wu Y, Li Y. Inflammation and tumor progression: signaling pathways and targeted intervention. Sign Transduct Targeted Ther. 2021 doi: 10.1038/s41392-021-00658-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
For the corresponding data and code, please contact the authors.