Skip to main content
Computational and Mathematical Methods in Medicine logoLink to Computational and Mathematical Methods in Medicine
. 2021 Aug 4;2021:7471516. doi: 10.1155/2021/7471516

Bayesian Gene Selection Based on Pathway Information and Network-Constrained Regularization

Ming Cao 1,2, Yue Fan 1, Qinke Peng 1,
PMCID: PMC8360753  PMID: 34394707

Abstract

High-throughput data make it possible to study expression levels of thousands of genes simultaneously under a particular condition. However, only few of the genes are discriminatively expressed. How to identify these biomarkers precisely is significant for disease diagnosis, prognosis, and therapy. Many studies utilized pathway information to identify the biomarkers. However, most of these studies only incorporate the group information while the pathway structural information is ignored. In this paper, we proposed a Bayesian gene selection with a network-constrained regularization method, which can incorporate the pathway structural information as priors to perform gene selection. All the priors are conjugated; thus, the parameters can be estimated effectively through Gibbs sampling. We present the application of our method on 6 microarray datasets, comparing with Bayesian Lasso, Bayesian Elastic Net, and Bayesian Fused Lasso. The results show that our method performs better than other Bayesian methods and pathway structural information can improve the result.

1. Introduction

Identifying disease-associated genes, which can be treated as diagnostic biomarkers, can bring a significant effect on disease diagnosis, prognosis, and treatments [1, 2]. With the development of high-throughput technologies in recent years, gene expression profiling has provided a useful way to find biomarkers. Researchers can identify the genes which are differentially expressed between two groups of samples. These genes are regarded as disease-associated genes. However, gene expression data usually contains a large number of genes and a relatively small sample size [3, 4]. And many of the genes are also redundant or irrelevant to the prediction [5, 6]. Furthermore, there are also noises in the experiment procedures which will influence the gene expression values. Therefore, identifying the biomarkers from gene expression data is challenging.

During the last decades, a number of gene selection methods have been developed to tackle this problem. Feature selection and feature extraction are two major methods (we treat gene and feature equally in this paper). On the one hand, the aim of feature selection is to select relevant features and do not change the form of the features. On the other hand, feature extraction will extract the feature from the original data and may alter the form of the features. Here, we focus on the feature selection methods since the results of such methods could be interpreted easily. Feature selection methods can be generally organized into three categories: filter, wrapper, and embedded methods. Both the wrapper and embedded methods are classifier-dependent methods; thus, they are always time consuming and easy to overfitting. However, the filter methods are usually based on statistic approaches [7] such as mRMR [5], PLSRFE [8], lasso [9], and elastic net [10], which are relatively efficient in terms of computation and can derive a score of each of the genes which represents the significance of the gene. Therefore, we focus on the filter methods in this paper.

Although these methods are successful in many applications, they usually obtain suboptimal solutions. Therefore, the prediction accuracies are not satisfied and the disease-associated genes selected from different methods have few overlaps [11]. This is partly due to the fact that the discriminatory power of many biomarkers is similar. Furthermore, some genes which have low discriminatory powers play important roles in cellular functions. Their combinations are highly discriminative while they are usually ignored [12, 13].

Recently, with a large amount of biological information accumulated, there is an increased interest in gene selection with incorporating information on pathways, which can partially compensate for the lack of reliable expression data [14]. Pathways depict a series of chemical interactions in living cells; genes that interact with one another usually mean that they function together concertedly. Therefore, these genes should be highly correlated and have dependence structures. However, many studies only utilize the information that pathways cluster genes into the natural group; the pathway structural information is neglected. Li and Li have overcome this disadvantage by incorporating pathway structure information through a Laplacian matrix of a global graph [15, 16] and combined with lasso penalty to perform network-constrained penalty which can select subgroups of correlated features in the network. This penalty is based on the assumption that genes belonging to the same pathway have similar functions and therefore smoothed regression coefficients. And this penalty has been successfully applied in many studies [1719].

The Bayesian approach has three major advantages over Bayesian selection methods [20]. Firstly, hyperparameters can be estimated automatically through fulfilling stochastic draws; thus, 10-fold cross-validation for estimating penalized parameters is not required. Secondly, the Bayesian framework can utilize the pathway information naturally by integrating it in the model as prior knowledge. Finally, the Bayesian estimation with the posterior distributions can provide credible intervals for the regression coefficients, which is a great advantage over frequentist methods.

In this paper, we work with a Bayesian framework to perform gene selection through network-constrained regularization. Similar to the Bayesian Lasso [21], Bayesian Elastic Net [22], and Bayesian Fused Lasso [23], we use shrinkage priors to perform regularization. We show that all the conditional posteriors of the proposed model are available in closed form and proper. Thus, parameter estimation can be performed through Gibbs sampling easily. The pathway information is obtained from the Kyoto Encyclopedia of Genes and Genomes (KEGG) [24], which is the most popular pathway public database, especially pathways associated with several types of cancer could be obtained in the model. Furthermore, following Held and Holmes [25], we extend the regression model to binary regression which can perform binary classification through an auxiliary variable. This method is assessed by applying it to several microarray datasets.

2. Method

2.1. The Bayesian Network-Constrained Model for Gene Selection

Considering an N × P matrix X, where P is the number of genes and N is the number of the samples, with a response vector y = (y1, ⋯,yn)T, we normalize the values of each feature as the tradition in variable selection; thus, the mean and standard deviation of each feature are 0 and 1. We assume the likelihood function of the continuous response is Gaussian function:

YX,β,σ2NnXβ,σ2In, (1)

which can be also expressed as

y=Xβ+ε,εNn0,σ2In. (2)

Following Li and Li's work [16], we incorporate the pathway information through its normalized Laplacian matrix. Consider an undirected graph G = (V, E, W). In this graph, genes are represented by a set of nodes V, and the interactions between genes are represented by a set of edges E = {u ~ v}, and W is the weights of the edges, where w(u, v) represents the weight of edge e = (u ~ v) which indicates the uncertainty of the edge between the vertices u and v. The degree of each vertex is defined as dv = ∑uvw(u, v). Then, the normalized Laplacian matrix L for graph G with the uth and vth elements can be defined by

Lu,v=1wu,vdu,if u=v and du0,wu,vdudv,if u and v are adjacent,0,otherwise. (3)

Here, we let w(u, v) = 1 if there exists an interaction between gene u and v, and w(u, v) = 0, otherwise.

To form the network-constrained regularization, we assign the prior distribution for β as follows:

βNp0,σ2rΛ1, (4)

where Λ is taking the form:

Λ=diagτ11,τ21,,τp1+L=1+τ11L1,2L1,pL2,11+τ21L2,pLp,1Lp,21+τp1. (5)

Note that Λ only contains hyperparameter τ.

To eliminate the |Λ|1/2 in the prior distribution of β, we assign the prior distribution for τ as follows:

pτ2λ=CτΛ1/2j=1pλ22expλ22τj2, (6)

where Cτ is the normalizing constant.

The prior distribution defined in (6) is proper, due to the following analysis:

Let A = ΛIn, and A is a symmetric and positive semidefinite matrix.

Let DA = diag(a1, ⋯, ap), where a1, ⋯, ap are eigenvalues of A and 0 ≤ a1 ≤ ⋯≤ap.

Since A is the symmetric and positive semidefinite, there exists an orthonormal matrix Q. Hence, the eigendecomposition of matrix A can be written as A = QDAQT.

Because of Λ = A + In = QDAQT + QQT = Q(DA + In)QT, so ∣Λ | = ∏i=1n(ai + 1) ≥ 1.

Then,

0CτΛ1/2j=1pλ22expλ22τj2dτ2Cτ0j=1pλ22expλ22τj2dτ2<, (7)

where the integrand is kernels of the gamma density that indicates the integral is finite. Therefore, the prior distribution is proper.

Since

βTΛβ=βTD1β+uvβuduβvdv20,D=diagτ12,τ22,,τp2, (8)

Λ is positive semidefinite.

The joint posterior distribution can be written as

pβ,λ,σ2,τ2,rX,Yσ2n/2expYXβ22σ2σ2p/2rp/2Λ1/2exprβD1β+rβTLβ2σ2Λ1/2λ22expλ22τ2prpσ2pλ. (9)

Integrating out τ2, we have

pβ,λ,σ2,rX,Y=pβ,λ,σ2,τ2,rX,Ypτ2dτ20σ2n/3expYXβ22σ2σ2p/2rp/2Λ1/2exprβD1β+rβTLβ2σ2Λ2/2λ22expλ22τ2prpσ2pλdτ20expYXβ2+rβTLβ2σ2exprβD1β2σ2λ22expλ22τ2dτ2. (10)

Applying the fact as follows to the above equation:

a2expaz=012πsexpz22sa22expa2s2ds,a>0, (11)

we have

pβ,λ,σ2,rX,Y0expYXβ2+rβTLβ2σ2exprβD1β2σ2λ22expλ22τ2dτ2=expYXβ2+rλβ+rβTLβ2σ2. (12)

Thus, maximizing the posterior distribution is equivalent to minimizing the following equation:

Lr,λ,β=yXβTyXβ+rλβ1+λβTLβ, (13)

which has the same regularization term as the method proposed in [19].

We assign the prior distribution for σ2 as follows:

σ2Inverse Gammaa,b. (14)

And we assign the following prior for the hyperparameters r and λ:

rGammac,d,λGammae,f. (15)

Then, the hierarchical Bayesian model is

YX,β,σ2NnXβ,σ2In,βσ2,τ2,rNp0,σ2rΛ1,τ2λΛ1/2λ22expλ22τ2,σ2Inverse Gammaa,b,rGammac,d,λGammae,f. (16)

2.2. Gibbs Sampling Method

The likelihood is

pyX,β,σ2σ2n/2expYXβTYXβ2σ2. (17)

According to the above hierarchical model and the likelihood, the joint posterior distribution on data is

pβ,σ2,τ2,λ2,rY,Xσ2n/2expyXβTyXβ2σ2σ2p/2rp/2Λ1/2exprβTΛβ2σ2Λ1/2j=1pλ22expλ22τj2σ2aexpbσ2rcexpdrλ2eexpfλ2. (18)

Due to the fact that all the prior distributions are conjugated, the full conditional posterior distributions for the parameters have closed forms.

pβ,σ2,τ2,r,Y,XexpYXβTYXβ2σ2exprβΛβ2σ2expXX+rΛβ22YXβ2σ2. (19)

Let μ = (XX + )−1XY, Σ = σ2(XX + )−1 , we have

βσ2,τ2,r,X,YNpμ,Σ, (20)
pσ2β,τ2,r,Y,Xσ2n/2expyXβTyXβ2σ2σ2p/2exprβTΛβ2σ2σ2aexpbσ2σ2n+p/2aexpyXβTyXβ+rβTΛβ2+b1σ2, (21)
σ2β,τ2,r,Y,XInverse Gamman+p2+a,yXβTyXβ+rβTΛβ2+b, (22)
pτ2β,σ2,λ2,rexprβTΛβ2σ2λ22expλ22τ2. (23)

This implies that τ2 follows a generalized inverse Gaussian distribution:

τj2β,r,σ2,λ2GIG12,λ2,rβj2σ2,j=1,2,,p, (24)
prβ,σ2,τ2rp/2exprβTΛβ2σ2rcexpdrrp/2+cexpβTΛβ2σ2+dr, (25)
rσ2,β,τ2Gammap2+c,βTΛβ2σ2+b, (26)
pλ2τ2j=1pλ22expλ22τj2λ2eexpfλ2λ2p+eexp12j=1pτj2+fλ2, (27)
λ2τ2Gammap+e,12j=1pτj2+f. (28)

The Gibbs sampling scheme iterates as follows:

  1. Update β by sampling from (20)

  2. Update σ2 by sampling form (22)

  3. Update τ2 by sampling from (24)

  4. Update r by sampling from (26)

  5. Update λ by sampling from (28)

2.3. The Binary Response Case

Binary data such as absence or presence or different types of a disease are often used as response variables in gene selection problems. To perform binary classification, we use probit regression using auxiliary variables. Then, the model can be represented as follows:

Pyi=1=Xiβ, (29)

where Xi is the ith sample and P(yi = 1) is the probability of yi = 1. Here, latent variables Z = (z1, z2, ⋯, zn) are defined as

zi=Xiβ+ε,εNn0,σ2In. (30)

Then, the full conditional posterior distribution for each zi is truncated normal:

ziβ,Xi,yiNXiβ,σ2Izi>0,yi=1,NXiβ,σ2Izi0,otherwise. (31)

And Z follows a multivariate truncated normal distribution:

pZβ,σ2,X,YNnXβ,σ2Ini=1nIAi, (32)
Ai=ZiZi>0,Yi=1,ZiZi0,Yi=0. (33)

Sampling from this distribution directly is difficult. We use the method proposed in [26] to sample this latent variable.

Then, the hierarchical Bayesian model is

ZX,Y,β,σ2NnXβ,σ2Ini=1nIAi,βσ2,τ2,rNp0,σ2rΛ1,τ2λΛ1/2λ22expλ22τ2,σ2Inverse Gammaa,b,rGammac,d,λGammae,f. (34)

To derive the Gibbs sampling scheme, we only need to replace Y with Z in the Gibbs sampling scheme defined in Section 2.2. And the latent variables Z are sampled from (32).

3. Results

3.1. Datasets and Preprocessing

To demonstrate the effectiveness of our methods, a regression microarray dataset and 5 real-life binary classification microarray datasets were tested in this paper, which are described as follows. The pathway information was obtained from the KEGG database.

A breast cancer dataset was used to predict the survival time of patients [27]. We used gene expression profiles of 76 patients. Each patient was measured with 24481 probes. 3592 genes were found in the KEGG database from this dataset. We used the logarithm of survival times of patients as the response variable in this dataset.

The other 5 binary classification microarray datasets are shown in Table 1. No. genes mean the genes we found both existing in the microarray dataset and KEGG pathway database.

Table 1.

Binary classification microarray datasets used.

Dataset name No. genes Samples P/N References
Leukemia 1883 72 47/25 [28]
DLBCL 2427 77 58/24 [29]
Prostate 3238 102 50/52 [30]
GSE412 3234 108 60/48 [31]
GSE4922 4476 204 70/134 [26]

Lastly, the gene expression values were normalized; thus, its mean and standard deviation are 0 and 1.

3.2. Parameter Settings

In the procedure of Bayesian network-constrained regularization, we recommend small values for a, b, c, d, e, f in (16) and we set these values to 0.01 in our experiments. The Gibbs sampling iteration was conducted 6000 times, and we chose the second half of the samples to estimate the regression parameters. The posterior estimates of all parameters were obtained through the posterior averages of the chains. For the classification problem, the classifiers were built by a support vector machine (SVM). In this paper, we used the radial basic function as the kernel function in SVM. And the regularization parameter and the kernel width parameter were optimized by a grid search approach. We used Libsvm [32] to model the SVM.

3.3. Results and Analysis

In this section, we will describe the results on 6 microarray gene expression datasets (Table 1) to evaluate the performance of the proposed method. Our method was compared with the other three Bayesian regularized regression methods, including Bayesian Lasso, Bayesian Elastic Net, and Bayesian Fused Lasso. A comprehensive review of these methods can be found in [23]. When L = I, which means we know nothing about the pathway structure, the Bayesian network-constrained regularization is equivalent to Bayesian Elastic Net. And when L = O, our method is equivalent to Bayesian Lasso. These three methods can also be extended to perform binary classification through an auxiliary variable. We also used Gibbs sampling to perform parameter estimation. Previous review [23] also shows that these three Bayesian methods' performances are similar to and in some cases better than the frequentist methods. Prediction mean square error was used to evaluate the performance on regression problem. Meanwhile, ACC and AUC were used as the evaluation criteria for binary classification problem. According to previous studies, the number of important genes is probably about 50 [28]; thus, we selected the top 50 genes based on the absolute value of their regression coefficient for the binary classification problem.

Figure 1 shows the performance of all the four methods on the regression microarray dataset. And the classification performances on the five binary classification microarray datasets are summarized in Table 2. In the binary classification datasets, the first three datasets are usually treated as easy classification datasets, while the other two datasets are relatively hard to classify. From Figure 1, we can see that the PMSE of our method is lower than other Bayesian methods. Table 2 also shows that on the four easy binary classification datasets, our method achieves the highest ACC and AUC. In the other two hard classification datasets, our method achieves the highest ACC and AUC on GSE412. Although the AUC of Bayesian Elastic Net is higher than our method on GSE4922, our method achieves the highest ACC. In general, Bayesian network-constrained regularization shows better prediction and classification ability than other three Bayesian methods, which is similar to the results implied by [15]. Since our method can be transferred to Bayesian Lasso or Bayesian Elastic Net when the normalized Laplacian matrix L = O or L = I, the results also show that pathway information indeed contributes to the accuracy of the gene selection.

Figure 1.

Figure 1

PMSE performance on regression microarray dataset.

Table 2.

Comparison of results of 4 Bayesian methods.

Dataset Methods AUC ACC
Leukemia BEN 0.9955 0.9600
BFused 1 0.9733
BLasso 1 0.9447
BNR 1 0.9733
DLBCL BEN 0.9674 0.9223
BFused 0.9674 0.9223
BLasso 0.9485 0.9223
BNR 0.9958 0.9482
Prostate BEN 0.9784 0.9414
BFused 0.9655 0.9314
BLasso 0.9784 0.9419
BNR 0.9900 0.9510
GSE412 BEN 0.9428 0.8498
BFused 0.9046 0.8619
BLasso 0.9541 0.8792
BNR 0.9637 0.9074
GSE4922 BEN 0.6274 0.6666
BFused 0.6028 0.6523
BLasso 0.6132 0.6860
BNR 0.6132 0.6860

Consistent with previous studies [33, 34], all the Bayesian regularization regression methods could classify Leukemia, DLBCL, Prostate, and GSE412 dataset accurately. However, the performances of all the methods were poor on GSE 4922 dataset. Therefore, we demonstrate the effectiveness of our method by selecting the top 18 genes which make the prediction accuracy to achieve the highest value and most of those genes are associated with breast cancer (Table 3).

Table 3.

Description of top 18 genes of GSE4922.

Gene symbol Description Reference
SYCP3∗ Synaptonemal complex protein 3 [35]
CDKN2A∗ Cyclin dependent kinase inhibitor 2A [36]
PLB1∗
CTNNBIP1∗
Phospholipase B1
Catenin beta-interacting protein 1
[37]
[38]
GBE1∗ 1,4-Alpha-glucan-branching enzyme 1 [39]
SMURF1∗ SMAD-specific E3 ubiquitin protein ligase 1 [40]
NR1H4 Nuclear receptor subfamily 1 group H member 4 /
PDE11A Phosphodiesterase 11A /
UGT1A1∗ UDP glucuronosyltransferase family 1 member A1 [41]
FGF19∗ Fibroblast growth factor 19 [42]
OR51B4∗ Olfactory receptor family 51 subfamily B member 4 [43]
RAB7A∗ RAB7A, member RAS oncogene family [44]
SDHD∗ Succinate dehydrogenase complex subunit D [45]
IFNA8 Interferon alpha 8 /
VANGL2∗ VANGL planar cell polarity protein 2 [46]
UMPS Uridine monophosphate synthetase /
CASP3 Caspase 3 [47]
SUFU SUFU negative regulator of hedgehog signaling [48]

The gene was reported as an oncogene in previous literatures.

4. Conclusion

In this paper, we propose a Bayesian approach to perform gene selection, which can incorporate the pathway information as prior biological knowledge through network-constrained regularization to improve the accuracy of gene selection. All the prior distributions we propose are strictly conjugated; thus, all the conditional posteriors of the model are available in closed form. An auxiliary variable is also introduced to extend the regression model to perform binary classification. An efficient Gibbs sampling method is used to estimate regression coefficients and tune parameters simultaneously, which can perform feature filter feasible for high dimensional microarray datasets. The performance of the proposed method is demonstrated by applying it to a regression microarray dataset and five binary classification microarray datasets. The results show that compared with Bayesian Lasso, Bayesian Elastic Net, and Bayesian Fused Lasso, our method performs better both in prediction and classification. And the pathway information indeed improves the accuracy of gene selection.

Acknowledgments

This work is supported in part by the National Science Foundation of China, under Grant 61173111.

Data Availability

The breast cancer dataset could be obtained from the R package breast cancer NKI. Leukemia, DLBCL, and Prostate datasets are available on the website http://portals.broadinstitute.org/cgi-bin/cancer/. GSE412 and GSE4922 datasets are available in the GEO of NCBI under accession GSE412 and GSE4922.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

References

  • 1.Govindarajan R., Duraiyan J., Kaliyappan K., Palanisamy M. Microarray and its applications. Journal of Pharmacy & Bioallied Sciences. 2012;4(Supplement 2):S310–S312. doi: 10.4103/0975-7406.100283. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Wang H., Jing X., Niu B. A discrete bacterial algorithm for feature selection in classification of microarray gene expression cancer data. Knowledge-Based Systems. 2017;126:8–19. doi: 10.1016/j.knosys.2017.04.004. [DOI] [Google Scholar]
  • 3.Dessì N., Pes B. Similarity of feature selection methods: an empirical study across data intensive classification tasks. Expert Systems with Applications. 2015;42(10):4632–4642. doi: 10.1016/j.eswa.2015.01.069. [DOI] [Google Scholar]
  • 4.Lu L., Townsend K. A., Daigle B. J., Jr. GEOlimma: differential expression analysis and feature selection using pre-existing microarray data. BMC Bioinformatics. 2021;22(1):p. 44. doi: 10.1186/s12859-020-03932-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Peng H., Long F., Ding C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2005;27(8):1226–1238. doi: 10.1109/TPAMI.2005.159. [DOI] [PubMed] [Google Scholar]
  • 6.Sun S., Peng Q., Zhang X. Global feature selection from microarray data using Lagrange multipliers. Knowledge-Based Systems. 2016;110:267–274. doi: 10.1016/j.knosys.2016.07.035. [DOI] [Google Scholar]
  • 7.Kanti Ghosh K., Begum S., Sardar A., et al. Theoretical and empirical analysis of filter ranking methods: experimental study on benchmark DNA microarray data. Expert Systems with Applications. 2021;169:p. 114485. doi: 10.1016/j.eswa.2020.114485. [DOI] [Google Scholar]
  • 8.You W., Yang Z., Ji G. Feature selection for high-dimensional multi-category data using PLS-based local recursive feature elimination. Expert Systems with Applications. 2014;41(4):1463–1475. doi: 10.1016/j.eswa.2013.08.043. [DOI] [Google Scholar]
  • 9.Tibshirani R. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological) 1996;58(1):267–288. doi: 10.1111/j.2517-6161.1996.tb02080.x. [DOI] [Google Scholar]
  • 10.Hui Z., Hastie T. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society. 2005;67(5):768–768. [Google Scholar]
  • 11.Fröhlich H. Network based consensus gene signatures for biomarker discovery in breast cancer. PLoS One. 2011;6(10, article e25364) doi: 10.1371/journal.pone.0025364. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Michailidis G. Statistical challenges in biological networks. Journal of Computational and Graphical Statistics. 2012;21(4):840–855. doi: 10.1080/10618600.2012.738614. [DOI] [Google Scholar]
  • 13.Wu M. Y., Zhang X. F., Dai D. Q., Ou-Yang L., Zhu Y., Yan H. Regularized logistic regression with network-based pairwise interaction for biomarker identification in breast cancer. BMC Bioinformatics. 2016;17(1):p. 108. doi: 10.1186/s12859-016-0951-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Su J., Yoon B. J., Dougherty E. R. Accurate and reliable cancer classification based on probabilistic inference of pathway activity. PLoS One. 2009;4(12, article e8161) doi: 10.1371/journal.pone.0008161. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Li L. H., Li H. Network-constrained regularization and variable selection for analysis of genomic data. Bioinformatics. 2008;24(9):1175–1182. doi: 10.1093/bioinformatics/btn081. [DOI] [PubMed] [Google Scholar]
  • 16.Li C., Li H. Variable selection and regression analysis for graph-structured covariates with an application to genomics. The Annals of Applied Statistics. 2010;4(3):1498–1516. doi: 10.1214/10-AOAS332. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Liu J., Huang J., Ma S. Incorporating network structure in integrative analysis of cancer prognosis data. Genetic Epidemiology. 2013;37(2):173–183. doi: 10.1002/gepi.21697. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Zhang W., Wan Y. W., Allen G. I., Pang K., Anderson M. L., Liu Z. Molecular pathway identification using biological network-regularized logistic models. BMC Genomics. 2013;14(Supplement 8):p. S7. doi: 10.1186/1471-2164-14-s8-s7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Jiang H., Hong T., Wang T., et al. Gene expression profiling of human bone marrow mesenchymal stem cells during osteogenic differentiation. Journal of Cellular Physiology. 2019;234(5):7070–7077. doi: 10.1002/jcp.27461. [DOI] [PubMed] [Google Scholar]
  • 20.Fan Y., Wang X., Peng Q. Inference of gene regulatory networks using Bayesian nonparametric regression and topology information. Computational and Mathematical Methods in Medicine. 2017;2017:8. doi: 10.1155/2017/8307530.8307530 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Hui Z. The Bayesian Lasso. 2008.
  • 22.Li Q., Lin N. The Bayesian elastic net. Bayesian Analysis. 2010;5(1):151–170. doi: 10.1214/10-ba506. [DOI] [Google Scholar]
  • 23.Casella G., Ghosh M., Gill J., Kyung M. Penalized regression, standard errors, and Bayesian lassos. Bayesian Analysis. 2010;5(2):369–412. doi: 10.1214/10-ba607. [DOI] [Google Scholar]
  • 24.Kanehisa M., Araki M., Goto S., et al. KEGG for linking genomes to life and the environment. Nucleic Acids Research. 2008;36(Database issue):D480–D484. doi: 10.1093/nar/gkm882. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Held L., Holmes C. C. Bayesian auxiliary variable models for binary and multinomial regression. Bayesian Analysis. 2006;1(1):145–168. doi: 10.1214/06-ba105. [DOI] [Google Scholar]
  • 26.Ivshina A. V., George J., Senko O., et al. Genetic reclassification of histologic grade delineates new clinical subtypes of breast cancer. Cancer Research. 2006;66(21):10292–10301. doi: 10.1158/0008-5472.CAN-05-4414. [DOI] [PubMed] [Google Scholar]
  • 27.van 't Veer L. J., Dai H., van de Vijver M. J., et al. Gene expression profiling predicts clinical outcome of breast cancer. Nature. 2002;415(6871):530–536. doi: 10.1038/415530a. [DOI] [PubMed] [Google Scholar]
  • 28.Golub T. R., Slonim D. K., Tamayo P., et al. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999;286(5439):531–537. doi: 10.1126/science.286.5439.531. [DOI] [PubMed] [Google Scholar]
  • 29.Shipp M. A., Ross K. N., Tamayo P., et al. Diffuse large B-cell lymphoma outcome prediction by gene-expression profiling and supervised machine learning. Nature Medicine. 2002;8(1):68–74. doi: 10.1038/nm0102-68. [DOI] [PubMed] [Google Scholar]
  • 30.Singh D., Febbo P. G., Ross K., et al. Gene expression correlates of clinical prostate cancer behavior. Cancer Cell. 2002;1(2):203–209. doi: 10.1016/S1535-6108(02)00030-2. [DOI] [PubMed] [Google Scholar]
  • 31.Cheok M. H., Yang W., Pui C. H., et al. Treatment-specific changes in gene expression discriminate _in vivo_ drug response in human leukemia cells. Nature Genetics. 2003;34(1):85–90. doi: 10.1038/ng1151. [DOI] [PubMed] [Google Scholar]
  • 32.Ferrari V. Libsvm : A Library for Support Vector Machines. 2008.
  • 33.Momenzadeh M., Sehhati M., Rabbani H. A novel feature selection method for microarray data classification based on hidden Markov model. Journal of Biomedical Informatics. 2019;95:p. 103213. doi: 10.1016/j.jbi.2019.103213. [DOI] [PubMed] [Google Scholar]
  • 34.Dabba A., Tari A., Meftali S., Mokhtari R. Gene selection and classification of microarray data method based on mutual information and moth flame algorithm. Expert Systems with Applications. 2021;166:p. 114012. doi: 10.1016/j.eswa.2020.114012. [DOI] [Google Scholar]
  • 35.Mobasheri M. B., Shirkoohi R., Modarressi M. H. Synaptonemal complex protein 3 transcript analysis in breast cancer. Iranian Journal of Public Health. 2016;45(12):1618–1624. [PMC free article] [PubMed] [Google Scholar]
  • 36.Lubiński J., Górski B., Huzarski T., et al. BRCA1-positive breast cancers in young women from Poland. Breast Cancer Research and Treatment. 2006;99(1):71–76. doi: 10.1007/s10549-006-9182-3. [DOI] [PubMed] [Google Scholar]
  • 37.Ueda A., Oikawa K., Fujita K., et al. Therapeutic potential of PLK1 inhibition in triple-negative breast cancer. Laboratory Investigation. 2019;99(9):1275–1286. doi: 10.1038/s41374-019-0247-4. [DOI] [PubMed] [Google Scholar]
  • 38.Mukherjee N., Dasgupta H., Bhattacharya R., et al. Frequent inactivation of MCC/CTNNBIP1 and overexpression of phospho-beta- cateninY654 are associated with breast carcinoma: clinical and prognostic significance. Biochimica et Biophysica Acta (BBA)-Molecular Basis of Disease. 2016;1862(9):1472–1484. doi: 10.1016/j.bbadis.2016.05.009. [DOI] [PubMed] [Google Scholar]
  • 39.Birch A. H., Quinn M. C. J., Filali-Mouhim A., Provencher D. M., Mes-Masson A. M., Tonin P. N. Transcriptome analysis of serous ovarian cancers identifies differentially expressed chromosome 3 genes. Molecular Carcinogenesis. 2008;47(1):56–65. doi: 10.1002/mc.20361. [DOI] [PubMed] [Google Scholar]
  • 40.Kwon A., Lee H. L., Woo K. M., Ryoo H. M., Baek J. H. SMURF1 plays a role in EGF-induced breast cancer cell migration and invasion. Molecules and Cells. 2013;36(6):548–555. doi: 10.1007/s10059-013-0233-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Kuo S.-H., Yang S. Y., You S. L., et al. Polymorphisms of ESR1, UGT1A1, HCN1, MAP3K1 and CYP2B6 are associated with the prognosis of hormone receptor-positive early breast cancer. Oncotarget. 2017;8(13):20925–20938. doi: 10.18632/oncotarget.14995. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Tiong K. H., Tan B. S., Choo H. L., et al. Fibroblast growth factor receptor 4 (FGFR4) and fibroblast growth factor 19 (FGF19) autocrine enhance breast cancer cells survival. Oncotarget. 2016;7(36):57633–57650. doi: 10.18632/oncotarget.9328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Zhang C., Zhao H., Li J., et al. The identification of specific methylation patterns across different cancers. PLoS One. 2015;10(3, article e0120361) doi: 10.1371/journal.pone.0120361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Xie J., Yan Y., Liu F., et al. Knockdown of Rab7a suppresses the proliferation, migration, and xenograft tumor growth of breast cancer cells. Bioscience Reports. 2019;39(2) doi: 10.1042/BSR20180480. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Yu W., He X., Ni Y., Ngeow J., Eng C. Cowden syndrome-associated germline SDHD variants alter PTEN nuclear translocation through SRC-induced PTEN oxidation. Human Molecular Genetics. 2015;24(1):142–153. doi: 10.1093/hmg/ddu425. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Hatakeyama J., Wald J. H., Printsev I., Ho H. Y. H., Carraway K. L. Vangl1 and Vangl2: planar cell polarity components with a developing role in cancer. Endocrine-Related Cancer. 2014;21(5):R345–R356. doi: 10.1530/ERC-14-0141. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Pu X., Storr S. J., Zhang Y., et al. Caspase-3 and caspase-8 expression in breast cancer: caspase-3 is associated with survival. Apoptosis. 2017;22(3):357–368. doi: 10.1007/s10495-016-1323-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Alimirah F., Peng X., Gupta A., et al. Crosstalk between the vitamin D receptor (VDR) and miR-214 in regulating SuFu, a hedgehog pathway inhibitor in breast cancer cells. Experimental Cell Research. 2016;349(1):15–22. doi: 10.1016/j.yexcr.2016.08.012. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The breast cancer dataset could be obtained from the R package breast cancer NKI. Leukemia, DLBCL, and Prostate datasets are available on the website http://portals.broadinstitute.org/cgi-bin/cancer/. GSE412 and GSE4922 datasets are available in the GEO of NCBI under accession GSE412 and GSE4922.


Articles from Computational and Mathematical Methods in Medicine are provided here courtesy of Wiley

RESOURCES