Abstract
In cancer research, high-throughput profiling has been extensively conducted. In recent studies, the integrative analysis of data on multiple cancer patient groups/subgroups has been conducted. Such analysis has the potential to reveal the genomic commonality as well as difference across groups/subgroups. However, in the existing literature, methods with a special attention to the genomic commonality and difference are very limited. In this study, a novel estimation and marker selection method based on the sparse boosting technique is developed to address the commonality/difference problem. In terms of technical innovation, a new penalty and computation of increments are introduced. The proposed method can also effectively accommodate the grouping structure of covariates. Simulation shows that it can outperform direct competitors under a wide spectrum of settings. The analysis of two TCGA (The Cancer Genome Atlas) datasets is conducted, showing that the proposed analysis can identify markers with important biological implications and have satisfactory prediction and stability.
Keywords: Integrative analysis, commonality and difference, sparse boosting, cancer genomics
Introduction
In cancer research, high-throughput profiling has been extensively conducted, generating a large amount of genomic data. A major goal of genomic studies is to identify markers associated with cancer outcomes/phenotypes. A vast literature has been published. In most of the existing studies especially the early ones, the focus has been on a homogenous set of cancer patients. With the accumulation of data and evidences, there has been a growing attention to the “relationships” across different/heterogeneous cancer patient groups (1; 2). Here different groups may have different cancer types, subtypes, or biomarker values. As a representative example, building on The Cancer Genome Atlas (TCGA) Research Network, the National Cancer Institute (NCI) initiated the Pan-cancer project to examine the commonality and differences across cancer types (3). Other examples are presented in (4; 5; 6; 7). There are two equally important aspects of such studies. The first aspect is difference. Cancer is highly heterogeneous. Different cancer types/subtypes are expected to have different genomic basis. Even patients with the same cancer type/subtype often have diverse genomic markers, which can be reflected in differences in biomarker values. The second aspect is commonality: some genetic mutations, epigenetic changes, pathways, and others have been identified as playing important roles in multiple cancers. Most cancers share similar fundamental properties such as metastasis.To comprehensively understand cancer, it is important to identify both genomic difference and commonality.
With heterogeneous cancer patient groups, the most straightforward approach is to conduct meta-analysis, under which each group is analyzed separately, and the results are pooled and compared (8). Cancer genomic studies usually have the “small sample size, high-dimensional measurement” characteristic, leading to unsatisfactory results for each individual group and hence overall meta-analysis (9; 10; 11). In the recent studies, integrative analysis has been proposed as a viable solution and shown to outperform meta-analysis. Under integrative analysis, the raw data from multiple groups (datasets) are pooled prior to analysis. The heterogeneity across groups is taken into account in model estimation and marker selection (12). In “standard” integrative analysis, there is a lack of attention to the relationships across groups/datasets. There are also a few integrative analysis approaches that pay special attention to the similarity across groups/datasets. For example, the contrasted penalization approach (13) applies a penalty to smooth the regression coefficients of the same covariates across multiple datasets. Penalization (14) and sparse boosting (15) techniques have been adopted to promote similarity in model sparsity structures. It is noted that the aforementioned and some other approaches can achieve similarity but not commonality. That is, the identified effects/models may have similar but not exactly identical magnitudes. In the literature, the most relevant is the analysis conducted in (16), which also considers commonality and difference and adopts the penalization technique.
In this study, data on heterogeneous cancer patient groups are taken in consideration. The goal, beyond the “standard” estimation and variable selection, is to identify the genomics commonality and difference across groups. Compared to the single-group analysis, the proposed analysis can be more informative by revealing the common mechanisms as well as group-specific cancer genomic characteristics. Advancing from meta-analysis and integrative analysis (including those techniques that promote similarity), the proposed analysis can directly obtain commonality, that is, the shared common effects can have exactly equal, as opposed to similar, estimates. In both high-dimensional and low-dimensional statistical learning, it has been observed that there is no dominatingly better technique. The proposed approach is based on sparse boosting, which is one of the most recent and highly effective boosting techniques (17). Published studies (18; 19) have shown that sparse boosting has certain unique advantages (and also possible disadvantages) compared to penalization and other high-dimensional techniques. It is thus of interest to develop the proposed method beyond the penalization one (16).
The rest of the article is organized as follows. In Section 2, we describe the data and model settings. The proposed sparse boosting approach is described in Section 3. In Section 4, we conduct simulation and comparison with the alternatives. The proposed approach is also applied to two TCGA datasets. Section 5 concludes this paper. Additional technical details and numerical results are provided in Appendix.
Data and model settings
Consider the integrative analysis of M independent datasets. Each dataset may correspond to a different type/subtype of cancer or a different subpopulation of the same type of cancer. Let Y1, Y2, … , YM be the response variables and X1, X2, … , XM be the genomic measurements. In our numerical study, we analyze gene expression data. It is noted that the proposed approach can also be applied to other types of omics measurements. To simplify notation, assume that M datasets measure the same set of covariates. With minor modifications, the proposed approach can also accommodate partially matched covariate sets. In dataset m(= 1, 2, … , M) with nm i.i.d. observations, Ym is associated with Xm via Ym ~ Φ(Xm βm), where is the p-vector of regression coefficients. Model Φ is assumed to have a known form. In numerical study, we consider continuous outcomes under linear regression (LR) models and right censored survival outcomes under accelerated failure time (AFT) models. Details on the model settings and estimation objective functions are provided in Appendix A. It is noted that, with minor modifications, the proposed approach can also accommodate other types of outcomes/models. We consider the setting with p > nm, m = 1, 2, … , M, which matches practical cancer genomic data.
A common limitation shared by many of the existing studies is that there is insufficient attention to the “interconnections” among genomic measurements. Specifically, genes form functional groups, with those in the same groups tending to function coordinately and those in different groups behaving differently. In our analysis, the p genomic measurements are assumed to belong to K non-overlapping groups, with p(k) in group k. Here the groups can be constructed functionally (for example, based on pathway or GO information) or statistically (for example, via clustering). Let denote the coefficient vector of covariates in group k in dataset m. If , then group k represents a commonality. In contrast, if there are datasets m1 and m2 with , then group k represents a difference. It is noted that when M > 2, it is possible to define partial commonality (difference). Details are provided below. Our goal is to properly discriminate commonality and difference in estimation and variable selection.
Methods
The proposed approach is based on the boosting technique. Boosting is a generic statistical learning technique that aggregates a series of weak learners into a strong one. A long array of published studies show that boosting has many desirable advantages, including flexibility, implementation simplicity, affordable computational cost, and satisfactory prediction performance (20; 21). Boosting can accommodate multiple types of weak learners. In this article, we consider regression models with linear covariate effects and linear functions of covariates as weak learners, to achieve simple interpretability.
Sparse boosting
In the analysis of cancer genomic data, the selection of relevant genomic measurements is as important as model building. With the ordinary boosting, variable selection can be achieved with early stopping. However, it has been suggested that the resulted models are not “sparse enough”. Sparse boosting (SBoost) has been designed to further achieve sparsity (17). It imposes penalty to promote the selection of sparser models. As SBoost is the technical basis of the proposed approach, it is presented below for the completeness of this article.
Consider the analysis of the mth dataset. Use Lm(·) to denote the loss function. For the LR and AFT models, its form is described in Appendix A. The SBoost algorithm proceeds as described in Algorithm 1.
The key difference between SBoost and ordinary boosting is that the SBoost’s objective function has two components. The first is the lack-of-fit measure, as in ordinary boosting. The second component explicitly penalizes model complexity and hence can promote sparsity. In the literature, multiple model complexity measures have been used in SBoost, including BIC, AIC, MDL (minimum description length), and others. In Algorithm 1, we use an BIC type measure as an example, which has been adopted
Algorithm 1.
SBoost
| Step 1: Initialization. Denote βm[t] as the estimate of βm in the tth iteration. Set t = 0. Initialize βm[t] = 0. |
| Step 2: Fit and update. t = t + 1. |
| Compute , where es is a p-vector with the sth component equal to 1 and others equal to 0, and es,j is its jth component. |
| Update . |
| Step 3: Iteration. Repeat Step 2 for T times. |
| Step 4: Stopping. At the tth iteration, compute . Select the optimal number of iterations as . |
| Step 5: Output. The regression coefficients are estimated as . The strong learner for dataset m is . |
in multiple published studies. ν is the step size. Published studies suggest that its value is not crucial as long as it is not too big. In our numerical analysis, ν is set to be 0.1 following Buhlmann and Yu (17).
Sparse boosting for multiple datasets
Consider the analysis of M independent datasets using the SBoost technique. The integrative sparse boosting (Int-SBoost) algorithm proceeds as described in Algorithm 2.
Algorithm 2.
Int-SBoost
| Step 1: Initialization. Set t = 0. Initialize βm[t] = 0, m = 1, 2, … , M. |
| Step 2: Fit and update. t = t + 1. For m = 1, 2, … , M, compute . |
| Update , where ν is the step size as in SBoost. |
| Step 3: Iteration. Repeat Step 2 for T times. |
| Step 4: Stopping. At the tth iteration, compute . |
| Select the optimal number of iterations as . |
| Step 5: Output. The regression coefficients are estimated as , m = 1, 2, … , M. The strong learner for dataset m is . |
As an integrative analysis method, Int-SBoost simultaneously takes multiple datasets into consideration. In each iteration, Int-SBoost applies sparse boosting to each dataset separately. It differs from SBoost in that, in deciding the stopping rule, all datasets are considered together. Specifically, all datasets have the same selected number of iterations, roughly corresponding to the same amount of regularization. It is noted that it is possible to design an Int-SBoost method to also take multiple datasets into consideration in Step 2.
Sparse boosting to identify commonality and difference across datasets
Although Int-SBoost can conduct the integrative analysis of multiple datasets, it does not have an explicit mechanism to identify commonality and difference across datasets. To this end, we propose the CD (commonality and difference)-SBoost algorithm as described in Algorithm 3.
Compared to the existing boosting based methods, including those described above, CD-SBoost approach has two major distinctions. The first is that a new penalty is developed to explicitly quantify difference across multiple datasets at the group level. Specifically, in , the numerator first counts the number of groups with identical regression coefficients in two distinct datasets, and then sums over all pairs of distinct datasets. The denominator is a normalization constant and counts the maximal number of identical groups across the M datasets. With tuning parameter λ, this penalty takes value in [0, λ]. It is minimized if all groups behave the same across M datasets (i.e., ), and is maximized if all groups behave differently. As such, this penalty can directly promote equal regression coefficients, i.e., commonality. It is also flexible enough to allow difference. The proposed analysis conducts group-based analysis: a whole group will be concluded as behaving same or differently across multiple datasets. Moreover, as a special case, if the analyzed datasets have a natural order, for example if they correspond to different cancer stages, then the penalty can be revised as .
The second distinction is the computation of increments γ = (γ1, γ2, … , γM). Different from the existing integrative boosting approaches which determine the increments separately for each dataset, CD-SBoost determines all increments simultaneously. Specifically, for covariate s, a set Γs is constructed which includes all candidate γ. If the regression coefficients of group ks, which covariate s belongs to, are identical in some datasets, say datasets m1, m2, … , ml (that is, ), some datasets in {m1, m2, … , ml} are allowed to have identical increments in coefficients, so as to preserve commonality in these datasets. In contrast, if for any two distinct datasets m1 and m2, the increments γm’s are determined separately for each dataset m, and are allowed to differ for different m. For simplicity, we only update the regression coefficients of an individual dataset at a time.
Tuning parameter selection
For the selection of λ, we consider the high-dimensional BIC (HDBIC) criterion (22). Specifically, consider M independent datasets with sample sizes n1, … , nM and under linear regression models. Denote dfm as the number of important covariates in dataset m and RSSm as
Algorithm 3.
CD-SBoost
| Step 1: Initialization. Set t = 0. Denote β = (β1, β2, … , βM), and β[t] as the estimate of β in the tth iteration. Initialize β[t] = 0. |
| Step 2: Fit and update. t = t + 1. |
| (I). For s = 1, 2, … , p, determine the candidate set Γs of γ = (γ1, γ2, … , γM). Denote ks as the group that covariate s belongs to. |
| Case (a): For any m1 ≠ m2, . Compute . Then . |
| Case (b): There are m1, m2, … , ml (l ≥ 2), such that . Compute , where A is a non-empty subset of {m1, … , ml}. Define . Then , where IA is a length-M vector with the components in set A equal to 1 and others equal to 0. |
| (II). Compute . |
| Here , where λ is the tuning parameter. Es,m is a p × M matrix with the (s, m)th element equals to 1 and others equal to 0. |
| Update , where the step size ν is set to be 0.1. |
| Step 3: Iteration. Repeat Step 2 for T times. |
| Step 4: Stopping. At the tth iteration, compute . |
| Select the number of iterations as . |
| Step 5: Output. The regression coefficients are estimated as . The strong learner for dataset m is , m = 1, 2, … , M. |
the residual sum of squares. The HDBIC is defined as
It can be defined for other data and model settings accordingly. Compared to the ordinary BIC, HDBIC may generate sparser models and hence is more suitable for high-dimensional data.
A small example
To better appreciate the working characteristics of CD-SBoost, we consider a small simulation example. Specifically, three independent datasets are simulated, each with 50 subjects. p = 200 covariates belong to four non-overlapping groups. Among them, one group (group 1) behaves the same across datasets (that is, commonality), two groups (groups 3 and 4) behave “partially the same” (that is, they behave the same in two out of three datasets), and one group (group 2) behaves differently across datasets (that is, difference). In each dataset, each group has two important covariates. All nonzero coefficients are equal to 1. The responses are generated from linear regression models with errors standard normally distributed. More details on the simulation setting are provided in the next section. Beyond the proposed CD-SBoost method, we also consider (1) Int-SBoost. This method conducts integrative sparse boosting with multiple datasets (Algorithm 2), but lacks an explicit mechanism to encourage commonality; (2) Sep-SBoost. This method conducts sparse boosting on each dataset separately (Algorithm 1), and then results are combined across datasets. This is a meta-analysis strategy, and there is no consideration of commonality/difference across datasets; and (3) Pool-SBoost. This method combines all datasets directly into one and applies sparse boosting (Algorithm 1). With this approach, all genomic measurements/groups are identified as commonality, and there is no difference identified across datasets. Estimation results for one simulation replicate are shown in Table 1. It is noted that for this specific example, Int-SBoost and Sep-SBoost generate identical estimates. In term of selection of important covariates, CD-SBoost outperforms Sep-SBoost/Int-SBoost by identifying fewer false positives and outperforms Pool-SBoost by identifying fewer false positives and more true positives as well. The estimation performance is evaluated by RMSE (see the next section for its definition). The RMSE values are 1.217 (CD-SBoost), 1.256 (Sep-SBoost/Int-SBoost), and 3.808 (Pool-SBoost). CD-SBoost has the lowest estimation error. More conclusive simulation results based on multiple replicates are presented in the next section.
Table 1.
Small example: estimation results of one simulation replicate. Each group of covariates is highlighted in grey.
| Cov | True |
CD-SBoost |
Int-SBoost(Sep-SBoost) |
Pool-SBoost |
||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Data1 | Data2 | Data3 | Data1 | Data2 | Data3 | Data1 | Data2 | Data3 | Data1 | Data2 | Data3 | |
| 1 | 1 | 1 | 1 | 0.967 | 0.967 | 0.967 | 0.806 | 1.015 | 1.224 | 1.085 | 1.085 | 1.085 |
| 2 | 1 | 1 | 1 | 1.108 | 1.108 | 1.108 | 1.242 | 1.128 | 0.879 | 1.289 | 1.289 | 1.289 |
| 16 | 0.563 | |||||||||||
| 31 | 1 | 0.659 | 0.918 | |||||||||
| 32 | 1 | 0.497 | 0.749 | |||||||||
| 33 | 1 | 0.686 | 0.863 | |||||||||
| 34 | 1 | 0.766 | 0.974 | 0.329 | 0.329 | 0.329 | ||||||
| 35 | 1 | 0.548 | 0.568 | 0.42 | 0.42 | 0.42 | ||||||
| 36 | 1 | 0.862 | 0.98 | |||||||||
| 46 | 0.361 | 0.361 | 0.361 | |||||||||
| 55 | 0.34 | 0.34 | 0.34 | |||||||||
| 62 | 1 | 1 | 0.805 | 0.805 | 1.05 | 0.711 | 0.698 | 0.698 | 0.698 | |||
| 63 | 1 | 1 | 0.878 | 0.878 | 0.816 | 1.042 | 0.712 | 0.712 | 0.712 | |||
| 64 | 1 | 0.59 | 0.591 | |||||||||
| 65 | 1 | 0.496 | 0.493 | |||||||||
| 78 | 0.348 | 0.348 | 0.348 | |||||||||
| 81 | 1 | 1 | 1.025 | 1.025 | 1.114 | 1.296 | 0.779 | 0.779 | 0.779 | |||
| 82 | 1 | 1 | 0.771 | 0.771 | 0.93 | 0.718 | ||||||
| 83 | 1 | 0.905 | 1.024 | 0.475 | 0.475 | 0.475 | ||||||
| 84 | 1 | 0.768 | 0.725 | |||||||||
Numerical studies
Simulation
We simulate three independent datasets, each with 200 subjects. 1,000 covariates with continuous distributions are generated to mimic gene expression data analyzed below. These covariates belong to 20 non-overlapping groups. The sizes of the groups range from 20 to 80 to better mimic practical data. The covariates have marginally normal distributions with means 0 and variances 1. Covariates within the same groups are more strongly correlated than those in different groups. More details on the correlation structure are provided in Appendix B. In each dataset, there are 40 important covariates, with two important covariates per group. Three scenarios for the important covariates/groups are considered: (1) full commonality, where the three datasets share the same set of important covariates and also the same regression coefficients; (2) partial commonality. Here two cases are considered: (a) datasets 1 and 2 have the same set of important covariates and the same regression coefficients, while dataset 3 has a different set of important covariates; and (b) datasets 2 and 3 have the same set of important covariates and the same regression coefficients, while dataset 1 has a different set of important covariates; and (3) no commonality, where none of the important covariates are shared by the three datasets. The twenty groups are randomly allocated into the above three scenarios with proportions ρf, ρp, and ρn.
We consider continuous data under the LR model and censored survival data under the AFT model. With censored survival data, the censoring times are generated independently from uniform distributions. The parameters of censoring distributions is adjusted so that the censoring rates are close to 25%. Under both models, the random errors are generated from N(0, σ2), with σ2 = 1 and 3 representing two noise levels. The nonzero regression coefficients are either set all equal to 0.5 or generated randomly. In the latter case, the nonzero regression coefficients in dataset 2 are generated from , those specific to dataset 1 are from , and those specific to dataset 3 are from . Overall, we have the following simulation settings: (S1) Regression coefficients are randomly generated, and σ2 = 1; (S2) Regression coefficients are randomly generated, and σ2 = 3; (S3) Regression coefficients are all equal to 0.5, and σ2 = 1; and (S4) Regression coefficients are all equal to 0.5, and σ2 = 3.
To better gauge performance of CD-SBoost, we compare with Int-SBoost, Sep-SBoost, and Pool-SBoost described above. In addition, two penalization (Lasso) based approaches are also considered: (1) Sep-Lasso, which is similar to Sep-Boost and applies Lasso penalization to each dataset; and (2) Pool-Lasso, which is similar to Pool-SBoost but uses Lasso for estimation.
When evaluating CD-SBoost and alternative approaches, we are the most interested in identification performance. Specifically, we are interested in whether an approach can identify groups behaving the same (i.e., commonality). This is measured using the TP (true positive) and FP (false positive) numbers. More specifically, when computing positive, for a group, we count the number of pairs of distinct datasets where this group behaves the same, and then sum over all groups. For three datasets with full commonality, partial commonality, and difference having proportions ρf, ρp and ρn, respectively, the true number of positives, denoted as Nig, equals 20(2ρf + ρp). CD-SBoost and alternatives conduct variable selection. We also use TP and FP to evaluate variable selection performance. Here the definitions are the same as in the literature. Estimation and prediction performance is also evaluated. Specifically, estimation performance is measured by ERMSE (estimation RMSE), which is defined as . Prediction performance is measured by PRMSE (prediction RMSE), which is defined as in an LR model, and as in an AFT model.
Summary statistics are computed based on 100 replicates. Results for setting S1 and the LR model are presented in Table 2. The rest of the results are presented in Appendix C. Simulation suggests that the proposed CD-SBoost can significantly outperform the alternatives in identifying commonality groups. Specifically, Sep-SBoost and Sep-Lasso can hardly identify commonality. Int-SBoost also does not have a mechanism for identifying commonality. Under Pool-SBoost and Pool-Lasso, all groups are forced to behave the same across datasets. As such, they perform well with the full commonality groups, but fail with the partial commonality and difference groups.
Table 2.
Simulation under the LR model, scenario S1. In each cell, mean(SD).
| Variable |
Group |
ERMSE | PRMSE | |||
|---|---|---|---|---|---|---|
| TP | FP | TP | FP | |||
| (ρf, ρp, ρn) = (0.8, 0.2, 0), Nig = 36 | ||||||
| CD-SBoost | 117.4 (0.5) | 2 (1.4) | 34.2 (1.3) | 0.1 (0.3) | 0.6 (0.2) | 23.5 (0.7) |
| Int-SBoost | 79.8 (6.8) | 40 (8.2) | 0.3 (0.7) | 0 (0) | 4.1 (0.3) | 39.5 (1.1) |
| Sep-SBoost | 79.6 (6.7) | 40 (8.2) | 0.3 (0.7) | 0 (0) | 4.2 (0.2) | 39.8 (0.5) |
| Pool-SBoost | 81.1 (6.6) | 31.1 (12) | 36 (0) | 4 (0) | 4.3 (0.3) | 60.8 (3.5) |
| Sep-Lasso | 95.5 (7.9) | 177.1 (34.6) | 0 (0) | 0 (0) | 3.9 (0.3) | 34.4 (5.6) |
| Pool-Lasso | 116.4 (1.1) | 106 (25.9) | 36 (0) | 4 (0) | 2.7 (0.1) | 37.1 (2.0) |
| (ρf, ρp, ρn) = (0.6, 0.2, 0.2), Nig = 28 | ||||||
| CD-SBoost | 108.4 (3.2) | 8.9 (3.1) | 25 (1.9) | 1 (0.8) | 1.6 (0.3) | 24.7 (0.7) |
| Int-SBoost | 81.3 (3.9) | 39.6 (6.7) | 0.1 (0.3) | 0 (0) | 4.1 (0.2) | 40 (0.8) |
| Sep-SBoost | 81.2 (3.9) | 39.5 (6.7) | 0.1 (0.3) | 0.2 (0.4) | 4.2 (0.2) | 40.6 (1.1) |
| Pool-SBoost | 70.8 (6.7) | 32.4 (8.9) | 28 (0) | 12 (0) | 4.9 (0.2) | 70.8 (3.8) |
| Sep-Lasso | 97.7 (8.6) | 185 (35.2) | 0 (0) | 0 (0) | 4 (0.4) | 32.5 (6.1) |
| Pool-Lasso | 102.1 (3.2) | 114.4 (32.2) | 28 (0) | 12 (0) | 4.2 (0.1) | 52.3 (3.5) |
| (ρf, ρp, ρn) = (0.1, 0.9, 0), Nig = 22 | ||||||
| CD-SBoost | 102.3 (5.4) | 17 (7.8) | 18.9 (1.9) | 0.5 (1.1) | 1.6 ( 0.6) | 24.7 (1.5) |
| Int-SBoost | 78.7 (6.3) | 41.2 (5.5) | 0.1 (0.3) | 0 (0) | 4.2 (0.1) | 40.5 (0.8) |
| Sep-SBoost | 78.7 (6.3) | 41.2 (5.5) | 0.1 (0.3) | 0.6 (1.1) | 4.3 (0.3) | 41.3 (1.1) |
| Pool-SBoost | 47.3 (3.8) | 43.6 (12.8) | 22 (0) | 18 (0) | 5.9 (0.2) | 82.5 (2.7) |
| Sep-Lasso | 94.7 (6.6) | 180.8 (29.8) | 0 (0) | 0 (0) | 3.9 (0.3) | 33.1 (5.3) |
| Pool-Lasso | 75.9 (11.9) | 91.4 (30.6) | 22 (0) | 18 (0) | 5.6 (0.2) | 71.2 (5.5) |
| (ρf, ρp, ρn) = (0.4, 0.1, 0.5), Nig = 18 | ||||||
| CD-SBoost | 95.8 (3.9) | 22.9 (5.4) | 13.7 (1.7) | 1.9 (1.4) | 2.2 (0.2) | 26.2 (0.7) |
| Int-SBoost | 79 (7.7) | 43 (8.8) | 0 (0) | 0 (0) | 4 (0.4) | 39.9 (1.2) |
| Sep-SBoost | 79 (7.7) | 43 (8.8) | 0 (0) | 0.5 (0.7) | 4.1 (0.4) | 39.8 (1.2) |
| Pool-SBoost | 51 (5) | 27.5 (6.4) | 18 (0) | 22 (0) | 5.5 (0.1) | 79.6 (2) |
| Sep-Lasso | 93.5 (8.9) | 168.1 (63.1) | 0 (0) | 0 (0) | 3.8 (0.5) | 21.3 (6.1) |
| Pool-Lasso | 72.5 (7.7) | 78.1 (27.2) | 18 (0) | 22 (0) | 5.4 (0.2) | 70 (5.1) |
| (ρf, ρp, ρn) = (0.2, 0.2, 0.6), Nig = 12 | ||||||
| CD-SBoost | 92.5 (5.6) | 29.2 (8.8) | 8.3 (2.5) | 2.5 (1.2) | 3.3 (0.6) | 27.2 (1) |
| Int-SBoost | 77.4 (6.9) | 42.2 (8.1) | 0.1 (0.3) | 0 (0) | 3.9 (0.4) | 39 (1.3) |
| Sep-SBoost | 77.3 (6.7) | 42.2 (8.1) | 0.1 (0.3) | 0.6 (1) | 4.1 (0.4) | 39 (1.3) |
| Pool-SBoost | 32.8 (4.1) | 26.9 (6.2) | 12 (0) | 28 (0) | 6.1 (0.1) | 87.7 (3) |
| Sep-Lasso | 91.1 (8.1) | 170.1 (36.6) | 0 (0) | 0.1 (0.2) | 3.9 (0.5) | 33.9 (7) |
| Pep-Lasso | 53.1 (11.1) | 69.7 (34.7) | 12 (0) | 28 (0) | 6 (0.2) | 79 (6.2) |
| (ρf, ρp, ρn) = (0.1, 0, 0.9), Nig = 4 | ||||||
| CD-SBoost | 78.1 (7) | 35 (11.1) | 2.1 (1) | 3.9 (2) | 4.2 (0.8) | 30.1 (1.8) |
| Int-SBoost | 76.3 (6.5) | 45.3 (8.4) | 0 (0) | 0 (0) | 4.4 (0.5) | 38.2 (1.4) |
| Sep-SBoost | 76.3 (6.6) | 45.3 (8.4) | 0 (0) | 0.4 (0.5) | 4.4 (0.5) | 38.4 (1.2) |
| Pool-SBoost | 20.2 (2.8) | 35 (3.9) | 4 (0) | 36 (0) | 6.7 (0.6) | 96 (2.9) |
| Sep-Lasso | 88.9 (9.7) | 159.9 (41.9) | 0 (0) | 0.2 (0.1) | 3.9 (0.6) | 33.1 (7) |
| Pool-Lasso | 18.6 (6.9) | 58.6 (24.4) | 4 (0) | 36 (0) | 6.6 (0.1) | 93.3 (6.8) |
In the identification of important variables, CD-SBoost is observed to have favorable performance. Since Sep-SBoost analyzes each dataset separately, its performance is not strongly affected by the commonality/difference across datasets. Under almost all settings, Sep-SBoost performs worse than CD-SBoost. When the three datasets have a low level of commonality, for example (ρf, ρp, ρn) = (0.1, 0, 0.9), Pool-SBoost has the worst performance, which is as expected. Its performance improves when the level of commonality increases, however, is still inferior to CD-SBoost. Int-SBoost does not show significant advantage over Sep-SBoost in terms of selection. Under all simulation scenarios, the two penalization methods have inferior performance. Specifically, they perform reasonably in terms of true positives but identify too many false positives. In the evaluation of estimation and prediction, the CD-SBoost method is again observed to have favorable performance. The overall observed patterns are similar across simulation settings.
Analysis of leukemia data
Acute myeloid leukemia (AML) is a type of blood and bone marrow cancer. Here data are downloaded from TCGA (https://portal.gdc.cancer.gov/projects/TCGA-LAML). In our analysis, of interest is the regulation of blast count, which is an important diagnostic and prognostic marker, by gene expressions. Data on 173 patients are available for analysis. In the literature, it has been suggested that older patients have disease biology different from that of the young (23). However, in some of the existing analyses, insufficient attention has been paid to this difference. The patients are divided into two groups: (1) Young patients who are younger than 50; and (2) Old patients who are 50 years old or older. The sample sizes are 54 and 119, respectively. For this specific example, it is reasonable to expect both commonality (as the two groups have the same cancer) and difference (with the across-age difference suggested in the literature).
In the original data, there are 20,532 gene expressions. In the KEGG pathway database obtained from the Broad Institute, 5,266 unique genes represent 186 pathways. In the leukemia dataset, 4,243 genes have pathway information and are further considered. This selection is due to consideration on interpretability. The number of leukemia-associated genes is expected to be small to moderate. To improve the reliability of analysis, we further conduct a screening and remove genes whose marginal associations with the response are less significant (p-values> 0.01). Overall, a total of 841 genes are included in analysis. To describe the coordination among genes, we first build a network with edge weights equal to the correlations between two genes. This fully connected network is then filtered into a sparse one using the method in (24), which only preserves edges that are statistically significant with respect to a null model for the local assignment of weights to edges. This sparse network is then partitioned into 24 non-overlapping groups using the Louvain method (25), a fast greedy-optimization-based community detection method.
With CD-SBoost, 22 genes, representing 8 groups, are identified in at least one dataset. Out of the identified gene groups, 4 are identified as commonality. The detailed estimation results are provided in Table 3. A quick literature search suggests that the findings are biologically meaningful. For example, CEP89 encodes protein Centrosomal Protein 89, which is found to be involved in mitochondrial metabolism and neuronal functions (26). Both CD27 and CDC25B have been found to play an important role in AML. CD27 protein is a member of the TNF-receptor superfamily, which regulates T cell generation and immunity. Studies have shown that CD27 is elevated in the sera of AML patients and associated with poor survival (27). CDC25B is a member of the CDC25 family of phosphatases. It is involved in cell cycle progression, chromatin modulation, and transcriptional regulation (28). CDC25 is a candidate prognostic marker and therapeutic target not only for AML but also for other types of cancers (29). Adenosine deaminase 1 (ADA) has been found to be a diagnostic and prognostic marker for patients with acute lymphoblastic leukemia (30). CSF3R is identified only in the older patients. It encodes a protein which is the receptor for colony stimulating factor 3 and works as a key regulator of granulocytes (31). Abnormal expressions of CSF3R have been found in severe congenital neutropenia, chronic neutrophilic leukemia, and other types of cancer (31).
Table 3.
Analysis of leukemia data using CD-SBoost: identified genes and estimates. Genes/groups identified as full commonality are highlighted in grey.
| Gene | Data 1 | Data 2 | Gene | Data 1 | Data 2 |
|---|---|---|---|---|---|
| CEP89 | 0.136 | 0.136 | C6ORF89 | −0.084 | −0.084 |
| DPYSL3 | 0.044 | 0.044 | |||
| APH1B | −0.185 | ||||
| ATXN1 | −0.067 | −0.123 | DIP2B | −0.227 | |
| CLTCL1 | −0.150 | −0.091 | C3ORF49 | 0.229 | |
| CMA1 | −0.223 | −0.052 | LOC158267 | −0.167 | |
| CSF3R | −0.174 | HSA6077 | 0.095 | ||
| ACTN2 | −0.135 | ||||
| ADA | 0.081 | 0.081 | |||
| CD27 | −0.065 | −0.175 | |||
| COL4A3 | −0.116 | CYB5R1 | −0.102 | −0.102 | |
| DSC1 | −0.091 | −0.065 | |||
| CDHR3 | −0.106 | C14ORF123 | −0.247 | ||
| CDC25B | −0.137 | ALDH9A1 | −0.146 | −0.090 |
The data is also analyzed using the alternatives described above. The summary comparison results are provided in Table 4 and Table 5. Overall, it is found that different approaches lead to different findings. Sep-SBoost and Int-SBoost identify the same set of 17 genes in 8 groups. Sep-Lasso identifies 169 genes in 23 groups. With these three methods, all groups (with nonzero effects) are identified as behaving differently across the two datasets. Pool-SBoost identifies 4 genes from 4 groups, and Pool-Lasso identifies 3 genes from 3 groups. Detailed estimation results using the boosting alternatives are provided in Appendix C, and those using the Lasso alternatives are provided in Supplementary materials.
Table 4.
Analysis of leukemia data: numbers of overlapping genes.
| CD-SBoost | Int-SBoost | Sep-SBoost | Pool-SBoost | Sep-Lasso | Pool-Lasso | |
|---|---|---|---|---|---|---|
| CD-SBoost | 22 | 7 | 7 | 3 | 14 | 1 |
| Int-SBoost | 17 | 17 | 2 | 12 | 1 | |
| Sep-SBoost | 17 | 2 | 12 | 1 | ||
| Pool-SBoost | 4 | 2 | 1 | |||
| Sep-Lasso | 169 | 2 | ||||
| Pool-Lasso | 3 |
Table 5.
Analysis of leukemia data: numbers of identified gene groups, commonalities and differences
| Groups | Commonalities | Differences | |
|---|---|---|---|
| CD-SBoost | 8 | 4 | 4 |
| Int-SBoost | 8 | 0 | 8 |
| Sep-SBoost | 8 | 0 | 8 |
| Pool-SBoost | 4 | 4 | 0 |
| Sep-Lasso | 23 | 0 | 23 |
| Pool-Lasso | 3 | 3 | 0 |
In practical data analysis, it is challenging to objectively determine which set of results is more sensible. To complement the identification/estimation analysis, we evaluate prediction and stability, which may provide support to the analysis results to a large extent: if the identification/estimation results are reliable, prediction is expected to be more accurate and stability to be higher. Specifically, each dataset is randomly split into a training and a testing set with sizes 3:1. Estimation is conducted using the training set and used to make prediction for subjects in the testing set. The prediction RMSE is then computed to evaluate prediction performance. This process is repeated 100 times. The average RMSE values are 6.159 (CD-SBoost), 6.551 (Int-SBoost), 6.620 (Sep-SBoost), 6.859 (Pool-SBoost), 9.605 (Sep-SLasso), and 6.418 (Pool-SLasso), respectively. With this random splitting approach, we also evaluate the stability of findings. Specifically, the Observed Occurrence Index (OOI) values*, which are the probabilities of a finding being repeated, are 0.788 (CD-SBoost), 0.762 (Int-SBoost), 0.760 (Sep-SBoost), 0.264 (Pool-SBoost), 0.721 (Sep-SLasso) and 0.173 (Pool-SLasso), respectively. Note that the overlapping genes identified by Sep-Lasso and CD-SBoost usually have rather high OOIs, which leads to not too much difference in their OOIs. Overall, the improved prediction and stability support the validity of our analysis.
Analysis of kidney cancer data
TCGA has three kidney cancer datasets, on kidney renal clear cell carcinoma (KIRC), kidney renal papillary cell carcinoma (KIRP), and chromophobe kidney cancer (KICH), respectively. In this analysis, the outcome of interest is overall survival. For each dataset,we exclude the subjects with neo-adjuvant therapy, missing in survival outcome and gene expressions. The final sample sizes are 65 with 9 deaths (KICH), 291 with 44 deaths (KIRP), and 517 with 165 deaths (KIRC), respectively. As all three datasets are on kidney cancer, commonality is expected. With the (high) heterogeneity of cancer, difference is also expected. Data processing is conducted in a similar manner as described above, which screens 2,192 genes out of 20,534. With the same approach as described above, these genes are divided into 40 non-overlapping groups.
CD-SBoost identifies 51 genes representing 12 groups. Among those groups, 4 are identified as full commonality, 4 are identified as partial commonality, and the rest are identified as difference. Detailed estimation results are provided in Table 6. It is again found that the identification results are meaningful. For example, gene NMB is identified for all three subtypes. It encodes protein Neuromedin B and has been found to be a growth factor in cancer cells (32). Inhibition of the Neuromedin B signaling is a possible new target in cancer treatment (33). VPS33B is identified in KIRC and a known tumor suppressor in liver cancer. Gene GALC is identified in KIRP and a prognostic marker associated with both tumor and distant metastasis (34). Gene CSNK2A1, which is only identified in KICH, is a cancer prognostic marker (35).
Table 6.
Analysis of kidney cancer data using CD-SBoost: identified genes and estimates. Genes/groups identified as full commonality are highlighted in grey. Those identified as partial commonality are highlighted in light grey.
| Gene | Data 1 | Data 2 | Data 3 | Gene | Data 1 | Data 2 | Data 3 |
|---|---|---|---|---|---|---|---|
| OPH | 0.129 | 0.129 | HYAL2 | 0.093 | |||
| ASRGL1 | 0.109 | 0.109 | IDH2 | 0.046 | 0.046 | ||
| ARHGEF6 | 0.063 | ||||||
| BDNFAS | 0.202 | KCNF1 | −0.131 | −0.131 | −0.131 | ||
| RUNX1IT1 | −0.741 | LOC154339 | −0.069 | −0.069 | −0.069 | ||
| CEP170 | 0.258 | NMB | −0.136 | −0.136 | −0.136 | ||
| CEP76 | −0.003 | ||||||
| CIB1 | 0.053 | LOC100422366 | −0.091 | −0.091 | |||
| CIRBP | −0.003 | PIP5K1A | −0.087 | ||||
| CLASP1 | 0.556 | PRR5ARHGAP8 | 0.249 | ||||
| CNNM2 | −0.105 | SLC25A26 | 0.059 | ||||
| CSNK2A1 | −0.052 | SMC3 | −0.072 | ||||
| ACHM1 | −0.112 | SPNS2 | 0.144 | 0.047 | |||
| FLJ21736 | 0.078 | SUCLG2 | 0.155 | ||||
| HSPC139 | −0.138 | TAC2 | 0.180 | ||||
| LOC101928965 | 0.102 | TARS2 | −0.140 | ||||
| LOC651293 | −0.224 | ||||||
| EBF1 | 0.091 | MIEF1 | 0.200 | ||||
| EFCAB3 | 0.168 | 0.168 | 0.168 | ||||
| TMPRSS9 | −0.101 | −0.101 | −0.101 | ||||
| FAIM2 | 0.192 | ||||||
| FIS1 | −0.279 | TXLNG | −0.162 | −0.162 | −0.162 | ||
| FKBP1A | −0.187 | ||||||
| FLJ16103 | 0.168 | VPS33B | 0.173 | ||||
| FLJ44635 | 0.216 | XRN2 | −0.064 | −0.162 | −0.162 | ||
| FUT3 | −0.434 | YAF2 | −0.188 | −0.188 | |||
| FUT6 | 0.177 | ||||||
| GALC | −0.184 | ZNF121 | −0.132 | ||||
| HES4 | −0.190 | ||||||
| HIGD1B | 0.176 | ||||||
| HIST1H2BM | −0.147 | ||||||
Data is also analyzed using the alternatives. Table 7 and Table 8 again suggest that different approaches lead to different findings. Detailed estimation results using the boosting based approaches are provided in Appendix C, and those using the Lasso based approaches are provided in Supplementary materials. In prediction evaluation, as the outcome variable is censored survival, the logrank statistic is used. The average logrank statistics are 10.322 (CD-SBoost), 9.287 (Int-SBoost), 8.987 (Sep-SBoost), 8.823 (Pool-SBoost), 9.453 (Sep-SLasso), and 8.077 (Pool-SLasso), respectively. In stability evaluation, the OOI values are 0.907 (CD-SBoost), 0.854 (Int-SBoost), 0.854 (Sep-SBoost), 0.635 (Pool-SBoost), 0.882 (Pool-Lasso), and 0.912 (Sep-Lasso). Overall, the proposed CD-SBoost is found to have competitive performance.
Table 7.
Analysis of kidney cancer data: numbers of overlapping genes.
| CD-SBoost | Int-SBoost | Sep-SBoost | Pool-SBoost | Sep-Lasso | Pool-Lasso | |
|---|---|---|---|---|---|---|
| CD-SBoost | 51 | 24 | 24 | 11 | 35 | 19 |
| Int-SBoost | 49 | 49 | 14 | 38 | 19 | |
| Sep-SBoost | 49 | 14 | 38 | 19 | ||
| Pool-SBoost | 23 | 16 | 12 | |||
| Sep-Lasso | 238 | 47 | ||||
| Pool-Lasso | 95 |
Table 8.
Analysis of kidney cancer data: numbers of identified gene groups, commonalities and differences
| Groups | Commonalities | Differences | |
|---|---|---|---|
| CD-SBoost | 12 | 4 | 8 |
| Int-SBoost | 16 | 0 | 16 |
| Sep-SBoost | 16 | 0 | 16 |
| Pool-SBoost | 16 | 16 | 0 |
| Sep-Lasso | 30 | 0 | 30 |
| Pool-Lasso | 31 | 31 | 0 |
Conclusion
With the fast accumulation of cancer omics data, there is a growing popularity of across-cancer analysis. This study has explicitly focused on the commonality and difference across cancer patient groups, which has been considered in very limited literature. Complementing the existing penalization and other studies, CD-SBoost, a novel sparse boosting approach, is developed. Under simpler settings, it has been shown that boosting, penalization, and other techniques have specific advantages and do not dominate each other. Considering the satisfactory properties of boosting under other settings, the proposed methodological research is worthwhile. Comparing with the existing approach, the proposed CD-SBoost introduces a new penalty to promote commonality and, meanwhile, allow for difference. The newly proposed penalty has some similarity with the one in (15). However, there are key differences. Specifically, the penalty in (15) is concerned with sparsity structure and cannot promote equal regression coefficients. A novel computation of increments is proposed, which determines all increments simultaneously. In addition, the proposed analysis is conducted at the group level. That is, a whole group can be identified as either commonality or difference. This has been designed to accommodate the coordination among genomic measurements. These innovations in boosting technique may also benefit studies.
Under a wide spectrum of simulation settings, CD-SBoost is observed to have favorable performance. In data analysis, it has found significant differences across cancer patient groups, which may suggest the limitation of the existing analyses that assume homogeneity. Overall, this study may provide a practically useful new venue for analyzing cancer omics data under heterogeneity.
Beyond those considered in this article, there are many other types of cancer outcomes/models. It will be of interest to extend the proposed analysis to other outcomes/models. In the proposed analysis, the patient groups are assumed to be pre-defined. It will be of interest to examine whether it is possible to couple the proposed approach with data-dependent patient group identification. Data analysis results may also worth further investigation.
Supplementary Material
Acknowledgments
Funding
This study was supported by National Natural Science Foundation of China (11605288, 71771211), Fund for building world-class universities (disciplines) of Renmin University of China, and National Institute of Health (CA204120, CA216017).
Footnotes
We, firstly, compute OOI of each gene, and average over the 15 highest OOIs.
References
- [1].Ricketts CJ, Cubas AAD, Fan HH et al. The cancer genome atlas comprehensive molecular characterization of renal cell carcinoma. Cell Reports 2018; 23(1): 313–326. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Berger AC, Korkut A, Kanchi RS et al. A comprehensive pan-cancer molecular study of gynecologic and breast cancers. Cancer Cell 2018; 33(4): 690–705. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Weinstein JN, Collisson EA, Mills GB et al. The cancer genome atlas pan-cancer analysis project. Nature Genetics 2013; 45(10): 1113–1120. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Omberg L, Ellrott K, Yuan Y et al. Enabling transparent and collaborative computational analysis of 12 tumor types within the cancer genome atlas. Nature Genetics 2013; 45(10): 1121–1126. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Liu J, Huang J, Zhang Y et al. Integrative analysis of prognosis data on multiple cancer subtypes. Biometrics 2014; 70(3): 480–488. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [6].Lawrence MS, Stojanov P, Mermel CH et al. Discovery and saturation analysis of cancer genes across 21 tumor types. Nature 2014; 505(7484): 495–501. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].Hoadley KA, Yau C, Hinoue T et al. Cell-of-origin patterns dominate the molecular classification of 10,000 tumors from 33 types of cancer. Cell 2018; 173(2): 291–304. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [8].Guerra R and Goldstein D. Meta-analysis and Combining Information in Genetics and Genomics. Boca Raton, FL: Chapman and Hall/CRC, 2009. [Google Scholar]
- [9].Zhu R, Zhao Q, Zhao H et al. Integrating multidimensional omics data for cancer outcome. Biostatistics 2012; 17(4): 605–618. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [10].Huang Y, Huang J, Shia B et al. Identification of cancer genomic markers via integrative sparse boosting. Biostatistics 2012; 13(3): 509–522. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [11].Liu J, Huang J and Ma S. Integrative analysis of cancer diagnosis studies with composite penalization. Scandinavian Journal of Statistics 2014; 41(1): 87–103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Zhao Q, Shi X, Huang J et al. Integrative analysis of ”-omics” data using penalty functions. Wiley Interdisciplinary Reviews: Computational Statistics 2015; 7(1): 99–108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Shi X, Liu J, Huang J et al. Integrative analysis of high-throughput cancer studies with contrasted penalization. Genetic Epidemiology 2014; 38(2): 144–151. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Huang Y, Zhang Q, Zhang S et al. Promoting similarity of sparsity structures in integrative analysis with penalization. Journal of the American Statistical Association 2016; 112(517): 342–350. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Huang Y, Liu J, Yi HD et al. Promoting similarity of model sparsity structures in integrative analysis of cancer genetic data. Statistics in Medicine 2017; 36(3): 509–559. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Sun Y, Jiang Y, Li Y et al. Identification of cancer omics commonality and difference via community fusion. Statistics in Medicine 2019; DOI: 10.1002/sim.8027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Buhlmann P and Yu B. Sparse boosting. Journal of Machine Learning Research 2006; 7(Jun): 1001–1024. [Google Scholar]
- [18].Ma S, Huang Y, Huang J et al. Gene network-based cancer prognosis analysis with sparse boosting. Genetics Research 2012; 94(4): 205–221. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [19].Yue M, Li J and Ma S. Sparse boosting for high-dimensional survival data with varying coefficients. Statistics in Medicine 2018; 37(5): 789–800. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Buhlmann P and Hothorn T. Boosting algorithms: regularization, prediction and model fitting (with discussion). Statistical Science 2007; 22(4): 477–505. [Google Scholar]
- [21].Buhlmann P and Yu B. Boosting. Wiley Interdisciplinary Reviews: Computational Statistics 2010; 2(1): 69–74. [Google Scholar]
- [22].Ing C and Lai T. A stepwise regression method and consistent model selection for high-dimensional sparse linear models. Statistica Sinica 2011; 21: 1473–1513. [Google Scholar]
- [23].Almeida AM and Ramos F. Acute myeloid leukemia in the older adults. Leukemia Research Reports 2016; 6: 1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Serrano MA, Boguna M and Vespignani A. Extracting the multiscale backbone of complex weighted networks. Proceedings of the National Academy of Sciences of the United States of America 2009; 106(16): 6483–6488. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [25].Blondel VD, Guillaume JL, Lambiotte R et al. Fast unfolding of communities in large networks. Journal of Statistical Mechanics: Theory and Experiment 2008; 2008(10): 155–168. [Google Scholar]
- [26].van Bon BWM, Oortveld MAW, Nijtmans LG et al. Cep89 is required for mitochondrial metabolism and neuronal function in man and fly. Human Molecular Genetics 2013; 22(15): 3138–3151. [DOI] [PubMed] [Google Scholar]
- [27].Chen D, Gerasimcik N, Camponeschi A et al. Cd27 expression and its association with clinical outcome in children and adults with pro-b acute lymphoblastic leukemia. Blood Cancer Journal 2017; 7(6): 575. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Perwitasari O, Torrecilhas AC, Yan X et al. Targeting cell division cycle 25 homolog b to regulate influenza virus replication. Journal of Virology 2013; 87(24): 13775–13784. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [29].Reikvam H, Tamburini J, Skrede S et al. Antileukaemic effect of pi3k-mtor inhibitors in acute myeloid leukaemia-gene expression profiles reveal cdc25b expression as determinate of pharmacological effect. British Journal of Haematology 2014; 164(2): 200–211. [DOI] [PubMed] [Google Scholar]
- [30].Ebrahimi-Rad M, Khatami S, Ansari S et al. Adenosine deaminase 1 as a biomarker for diagnosis and monitoring of patients with acute lymphoblastic leukemia. Journal of Medical Biochemistry 2017; 37(2): 1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [31].Pankaj D and Kenneth DG. Granulocyte colony-stimulating factor receptor signaling in severe congenital neutropenia, chronic neutrophilic leukemia, and related malignancies. Experimental Hematology 2017; 46: 9–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].Gajjar S and Patel BM. Neuromedin: An insight into its types, receptors and therapeutic opportunities. Pharmacological Reports 2017; 69(3): 438–447. [DOI] [PubMed] [Google Scholar]
- [33].Moreno P, Mantey SA, Lee SH et al. A possible new target in lung-cancer cells: The orphan receptor, bombesin receptor subtype-3. Peptides 2018; 101: 213–226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [34].Liu DG, Xue L, Li J et al. Epithelial-mesenchymal transition and galc expression of circulating tumor cells indicate metastasis and poor prognosis in non-small cell lung cancer. Cancer Biomarkers 2018; (Preprint): 1–10. [DOI] [PubMed] [Google Scholar]
- [35].Rabjerg M. Identification and validation of novel prognostic markers in renal cell carcinoma. Danish medical journal 2017; 64(10). [PubMed] [Google Scholar]
- [36].Stute W Consistent estimation under random censorship when covariables are present. Journal of Multivariate Analysis 1993; 45(1): 89–103. [Google Scholar]
- [37].Karrer B and Newman MEJ. Stochastic blockmodels and community structure in networks. Physical Review E 2011; 83(2): 016107. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
