Skip to main content
Frontiers in Genetics logoLink to Frontiers in Genetics
. 2021 Oct 15;12:768747. doi: 10.3389/fgene.2021.768747

Hypertension-Related Drug Activity Identification Based on Novel Ensemble Method

Bin Yang 1, Wenzheng Bao 2,*, Jinglong Wang 3
PMCID: PMC8554208  PMID: 34721551

Abstract

Hypertension is a chronic disease and major risk factor for cardiovascular and cerebrovascular diseases that often leads to damage to target organs. The prevention and treatment of hypertension is crucially important for human health. In this paper, a novel ensemble method based on a flexible neural tree (FNT) is proposed to identify hypertension-related active compounds. In the ensemble method, the base classifiers are Multi-Grained Cascade Forest (gcForest), support vector machines (SVM), random forest (RF), AdaBoost, decision tree (DT), Gradient Boosting Decision Tree (GBDT), KNN, logical regression, and naïve Bayes (NB). The classification results of nine classifiers are utilized as the input vector of FNT, which is utilized as a nonlinear ensemble method to identify hypertension-related drug compounds. The experiment data are extracted from hypertension-unrelated and hypertension-related compounds collected from the up-to-date literature. The results reveal that our proposed ensemble method performs better than other single classifiers in terms of ROC curve, AUC, TPR, FRP, Precision, Specificity, and F1. Our proposed method is also compared with the averaged and voting ensemble methods. The results reveal that our method could identify hypertension-related compounds more accurately than two classical ensemble methods.

Keywords: hypertension, flexible neural tree, ensemble, network pharmacology, machine learning

Introduction

Hypertensive disease is a frequent cardiovascular disease characterized by elevated arterial blood pressure and accompanied by the target organ injury or clinical diseases (Essiarab et al., 2011; Owlia and Bangalore, 2016). It is a risk factor leading to many serious complications such as stroke, hypertensive heart disease, renal failure, atherosclerosis, and so on (Sakai and Sigmund, 2005; Brinks and Eckhart, 2010). Due to the increasing pressure of work and life, many people do not develop good eating and living habits, and often stay up late. The age of hypertensive patients tends to be younger. Therefore, the prevention and treatment of hypertension has become very important for human health.

Network pharmacology (NP) could construct a multi-dimensional network based on “traditional Chinese medicine prescription-chemical component-targets-disease targets” to analyze the relationships between traditional Chinese medicine multi-components and activity, which could provide a theoretical basis for further experimental research on a pharmacodynamic material basis and action mechanism (Wang et al., 2018; Xu et al., 2018). In recent years, network pharmacology has revealed therapeutic targets for hypertension and become a research hotspot, as it has been clinically verified to be an effective method of drug screening (Chen et al., 2020). Chen et al. screened out the key compounds and targets of JiaWeiSiWu granule to reveal the mechanism of JiaWeiSiWu granule in treating hypertension by NP method (Chen et al., 2021a). By NP and molecular docking (MD) methods, Zhai et al. investigated the mechanism of Pinellia ternate in treating hypertension (Zhai et al., 2021). Chen et al. analyzed the network based on Guizhi decoction, active compounds, and targets, and found hypertension-related targets and key pathways (Chen et al., 2021b). Chen et al. utilized NP and MD to analyzed the genistein for treating pulmonary hypertension (PH) and provided new guidance for further PH-related research (Chen et al., 2019). Liu et al. explained the pharmacological mechanism of TaohongSiwu decoction in the treatment of essential hypertension (EH) by the NP method (Liu et al., 2020). Wang et al. utilized NP to analyze the mechanism of Yeju Jiangya decoction against hypertension (Wang et al., 2021).

In recent decades, many data mining methods have been applied to reveal the disease mechanism and medication law of many complex diseases, especially hypertension (Ji and Wang, 2014; Ji et al., 2015; Hwang et al., 2016; Hu et al., 2018; Liang et al., 2018; Amaratunga et al., 2020; Liu et al., 2021; Zhao et al., 2021). Zhang et al. utilized SPSS21.0 and Apriori algorithm to analyze the symptom/sign information of EH patients collected and gave their distribution law and correlation (Zhang et al., 2019a). Yuan and Chen proposed niche technology and an artificial bee colony algorithm to mine association rules from Traditional Chinese Medicine (TCM) cases for treating hypertension (Yuan and Chen, 2011). Ma et al. collected the new literature about hypertension and constructed the gene network by analysis (Ma et al., 2018). Ramezankhani et al. utilized a decision tree to predict the risk factors of hypertension incidence in data collected from Iranian adults (Ramezankhani et al., 2016). Aljumah et al. utilized a data mining method to predict the treatment of hypertension patients with different age groups (Aljumah et al., 2011). Fang et al. proposed a new model-based KNN and LightGBM to predict the risk of hypertension (Fang et al., 2021).

Few studies have involved the use of data mining methods to improve network pharmacology. In this paper, a novel ensemble method based on a flexible neural tree (FNT) is proposed to identify hypertension-related active compounds. In the ensemble method, the used base classifiers are Multi-Grained Cascade Forest, support vector machines, random forest, AdaBoost, decision tree, Gradient Boosting Decision Tree, KNN, logical regression, and naïve Bayes. The classification results of nine classifiers are input to the FNT model, which is trained to predict hypertension-related compounds. The data used in the experiment are from up-to-date literature collected about hypertension and network pharmacology. By analysis of the literature, hypertension-related compounds were collected as positive samples and the generated decoys were utilized as negative samples. The molecular descriptor of each compound is extracted as the feature vector.

Methods

Classifiers

Assume that the training data is T={(x1,y1),(x2,y2),,(xn,yn)} containing n sample points. Sample point xi={xi1,xi2,xim} contains m features and category label yi={c1,c2} contains two cases. The nine classifiers used are introduced in the following sections of the article.

Multi-Grained Cascade Forest

Multi-Grained Cascade Forest (gcForest) is a novel ensemble machine learning method, which utilizes the cascade forest (ensemble of decision trees) to learn and generate models (Zhou and Feng, 2017). The core of gcForest mainly includes two modules: multi-grained scanning and cascade forest. The flowchart of gcForest is depicted in Figure 1.

  • 1) Multi-grained scanning

FIGURE 1.

FIGURE 1

The process of gcForest.

Multi granularity scanning is a technical means to enhance cascade forest and do more processing on features. Firstly, a complete m - dimensional sample is input, and then sliding sampling is carried out through the k1 -dimensional and k2 -dimensional sampling windows in order to obtain s1=(mk1)+1 and s2=(mk2)+1 feature subsample vectors, respectively. Each sub-sample is used for the training of completely random forest ( A ) and random forest ( B ). A probability vector with 2-dimension is obtained in each forest, so that two kinds of forests can produce 2s1 and 2s2 representation vectors, respectively. Finally, the results of all forests are spliced together to obtain the sample output.

  • 2) Cascade forest

Cascade forest includes several layers, each layer is composed of many forests, and each forest is composed of many decision trees. Completely random forest ( A ) and random forest ( B ) in each layer ensure the diversity of the model. For a completely random forest, each tree in the forest randomly selects a feature as the splitting node of the splitting tree, which grows until each leaf node is subdivided into only one class. For random forest, each tree randomly selects m candidate features, and the splitting nodes are filtered through the Gini coefficient. Each forest could generate a two-dimensional class vector. The two-dimensional class vectors of all forests are averaged to obtain the final two-dimensional class vector. Finally, the category with the maximum value in the final two-dimensional class vector is taken as the final classification result.

Support Vector Machines

Support vector machines (SVM) is a supervised learning algorithm based on statistical learning theory (Suykens and Vandewalle, 1999; Furey et al., 2000). With the sample set containing positive and negative samples, SVM could search a hyperplane that could segment the samples according to positive and negative classes. The classification hyperplane can be given as follows.

wTx+b=0. (1)

Where x is the data point on the classification hyperplane, w is a vector perpendicular to the classification hyperplane, and b is the displacement.

Linear separated data can be distinguished by the optimal classification hyperplane. For non-linear separated data, SVM can be transformed into solving the following optimization problem by the soft interval optimization and kernel techniques.

{minϕ(w,ς)=w2+12Ci=1nςis.t.yi[(wxi+b)]1ςi. (2)

Where C is the penalty factor, ςi is the relaxation variable, and xi is mapped to a high-dimensional space by ϕ . SVM could find a hyperplane with the largest interval in this high-dimensional space to classify the data.

Random Forest

Random forest (RF) is a machine learning method based on an ensemble of decision trees for classification and regression (Breiman, 2001; Díaz-Uriarte and Alvarez de Andrés, 2006). Random forest is a combined classification model composed of many decision tree classification models. Each decision tree has the right to vote to determine the best classification result. In random forest, firstly, K sample sets are extracted from the original training set by bootstrap sampling method, and the size of each extracted sample set is the same as that of the original training set. Then, K decision tree models are established from K sample sets, respectively. And K trees will create K classification results. The random forest integrates all the classified results by voting method, and the category with the most votes is designated as the final classification result.

AdaBoost

AdaBoost is a dynamic ensemble classification algorithm, which is to reasonably combine multiple weak classifiers (single-layer decision tree) to make it a strong classifier (Morra et al., 2009; Cao et al., 2013). The detailed algorithm is given as follows.

  • 1) Initialize the weight of each sample. Assuming that the dataset contains n samples, each training sample point is given the same weight ( 1n ) at the beginning.

  • 2) Train weak classifiers. According to the samples, the weak classifiers are trained. If a sample has been accurately classified, its weight will be reduced in constructing the next training set. On the contrary, if a sample point is not accurately classified, its weight is increased. At the same time, according to the classification error of the weak classifier, its weight is calculated. Then, the sample set with updated weights is used to train the next classifier, and the whole training process goes on iteratively. T weak classifiers are obtained after T iterations.

  • 3) The trained weak classifiers are combined into strong classifiers. Each weak classifier connects its respective weights through the classification function to form a strong classifier. After the training process of each weak classifier, the weight of the weak classifier with a smaller classification error rate is larger, which plays a greater decisive role in the final classification function, while the weight of the weak classifier with a larger classification error rate is smaller, which plays a smaller decisive role in the final classification function.

Decision Tree

A Decision Tree (DT) learning algorithm is usually a process of recursively selecting the optimal features and segmenting the training data according to the features so that each sub dataset has the best classification. The CART algorithm is one of the most common decision tree algorithms, which is mainly used for classification and regression (Breiman et al., 1984; Temkin et al., 1995). CART introduces the knowledge of probability theory and statistics into the research of decision tree. Different from the C4.5 algorithm, the CART algorithm could make a binary partition of the feature space and can split scalar attributes and continuous attributes. The specific algorithm is as follows:

  • 1) Calculate the Gini index of the existing features. The feature with the smallest Gini index is selected as the splitting attribute of the root node. According to the optimal feature and cut point, two sub-nodes are generated from the current node, and the training dataset is allocated to the two sub-nodes according to the feature. According to an attribute value, a node is segmented to make the data in each descendant subset more “pure” than the data in its parent subset. Gini coefficient measures the impurity of sample division, and the smaller the impurity is, the higher the “purity” of the samples is.

For 2-class problems, the training set S is divided into two subsets S1 and S2 according to an attribute A . The Gini coefficient of the given division S is calculated as follows.

GiniA(S)=|S1||S|Gini(S1)+|S2||S|Gini(S2). (3)

Where |S| is the number of samples in set S , and Gini(S)i is the Gini coefficient of sample set Si , which is calculated as follows:

Gini(Si)=1k=12(|Ck||Si|)2. (4)

Where |Ck| denotes the number of samples belonging to class k in the set Si .

  • 2) Step (1) is called recursively for two child nodes, and the iteration continues until the samples in all child nodes belong to the same category or no attributes can be selected as splitting attributes.

  • 4) Prune the CART decision tree generated.

Gradient Boosting Decision Tree

Gradient Boosting Decision Tree (GBDT) is an integrated learning algorithm (Hu and Min, 2018; Zhang et al., 2019b). By boosting method, N weak learners are created, which are combined into a strong learner after many iterations. The performance of the strong learner is higher than any weak learner. In GBDT, the used weak learner is the CART regression tree. During each iteration of GBDT, the residual of the previous model is reduced, and a new model is trained and established in the gradient direction of residual reduction, to improve the performance of the classifier. The specific algorithm is shown as follows:

  • 1) Initialize the weak learner.

f0(x)=argminκi=1nL(yi,κ). (5)

Where L is the loss function.

  • 2) For tth iteration ( t=1,2,,T )

  • a) For ith sample, the residual reduction is calculated as follows.

rti=[L(yi,f(xi))f(xi)]f(x)=ft1(x). (6)

Where ft1(x) is the classifier during the t1th iteration.

κtj=argminκxiRtjL(yi,ft1(xi)+κ). (7)

Where κtj is the value of the leaf node in the regression tree.

  • b) The calculated residues are used as new sample data, (xi,rti) is utilized to fit a new CART regression tree and the probability of each category is calculated. The leaf node region of the CART regression tree Rtj ( j=1,2,,J ) is obtained. J is the number of leaf nodes of the regression tree.

  • c) Calculate the optimal coefficient for the leaf area, which is given as follows.

  • d) The strong learner is updated with Eq. 8.
    ft(x)=ft1(xi)+j=1JκtjI(xRtj). (8)

When xRtj is true, I is equal to 1; otherwise, it is equal to 0.

  • 3) The final strong learner f(x) is obtained with Eq. 9.

f(x)=f0(x)+t=1Tj=1JctjI(xRtj). (9)

K-Nearest Neighbor

K-Nearest Neighbor (KNN) is a classification algorithm based on supervised learning, which is to classify the data points according to the sample set with the known categories (Liao and Vemuri, 2002). Select the K neighbors with the smallest distance from the input data in the training set, and take the category with the most times among the K neighbors as the category of the classified data point. In the KNN algorithm, the selected neighbors are objects that have been correctly classified.

In the KNN method, the most commonly used measurement of distance is the Euclidean distance. The Euclidean distance of two variables ( xi and xj ) is defined as follows.

D((xi,xj)=k=1m(xikxjk)2. (10)

Logistic Regression

Logistic regression (LR) is utilized to deal with the regression problem, which obtains the minimum result of cost function by gradient descent method to obtain the better classification boundary (Maalouf, 2011; Munshi et al., 2014). LR maps the values of linear regression to the interval [0, 1] by Sigmoid function, which is defined as follows.

yi=hθ(xi)=11+eθTxi. (11)

Where θTxi=θ0+θ1xi1+θ1xi2++θmxim , θ0 is a deviation parameter and θi represents the weight.

In order to solve the logistic regression model, the gradient descent algorithm is generally used to iteratively calculate the optimal parameters of the model.

Naïve Bayes

Naïve Bayes (NB) is one of the most widely utilized models in Bayesian classifiers, which is based on the assumption that the influence of an attribute value on the given class is independent of the values of other attributes (class conditional independence) (Rish, 2001; Li and Guo, 2005). The specific algorithm idea is as follows.

According to the joint probability and the prediction data x , the prediction category of x is defined as follows.

argmaxp(y=ck|x). (12)

According to the Bayesian theorem, p(y=ck|x) is calculated as follows.

p(y=ck|x)=p(x|y=ck)p(y=ck)p(x). (13)

Since the denominator is constant for all categories, just maximize the numerator, and Eq. 12 could be defined as follows.

argmaxp(x|y=ck)p(y=ck). (14)

Because each feature attribute is conditionally independent, p(x|y=ck) could be calculated as follows.

p(x|y=ck)=i=1mp(xi|y=ck) (15)

According to Eq. 15, Eq. 14 can be calculated as follows.

argmaxp(y=ck)i=1mp(xi|y=ck) (16)

Select the category with the largest posteriori probability as the prediction category.

Ensemble Methods

To improve the classification performance of a single classifier, a novel ensemble method based on a flexible neural tree (FNT) is proposed. An example of our proposed ensemble method is depicted in Figure 2. From Figure 2, it could be seen that the used base classifiers are gcForest, SVM, RF, AdaBoost, decision tree, GBDT, KNN, logical regression, and naïve Baye, which are introduced in detail in Classifiers. Firstly according to the training data, these nine classifiers can output their corresponding confidence level set ( c=(c1,c2,,c9) ), which is utilized as the input layer of the FNT model. The other hidden layers of the FNT model can be created randomly from operator set ( F=(+2,+3,,+n) ) and variable set ( T=(c1,c2,,c9) ) (Chen et al., 2006). +i denotes a flexible neuron operator, which can be calculated as follows:

{neti=j=1iwjxj,oi=f(ai,bi,neti)=e(netiaibi)2. (17)

Where f() is an activation function, ai and bi are the parameters of function, xj is the input variable and wj is the corresponding weight of the input variable.

FIGURE 2.

FIGURE 2

The flowchart of our proposed ensemble method.

FNT is a kind of cross-layer neural network, so each hidden layer can contain both operator and variable nodes. Because the structure of the FNT model is not fixed and this model contains many parameters such as ai , bi , and wj , many swarm algorithms have been proposed to search the optimal FNT model by iterations. In this paper, a hybrid evolutionary method based on genetic programming like structure optimization algorithm and simulated annealing was utilized for the training dataset. The detailed algorithms were introduced in another study (Yang et al., 2013).

Hypertension-Related Activity Drug Identification

In order to identify hypertension-related active compounds accurately, an ensemble method based on nine classifiers and a flexible neural tree is proposed. The process of hypertension-related active compounds identification is depicted in Figure 3. A total of 44 important studies were collected by querying the literature database according to two keywords: hypertension and network pharmacology. Through analyzing this literature, many important medicines such as Banxia Baizhu Tianma Tang, Chaihu Longgu Muli Decoction, compound reserpine and triamterene tablets, and Huanglian Jiedu Decoction, were collected and 88 hypertension-related compounds were searched. These important compounds were verified by biology experiments or molecular docking, which were used as positive samples in this paper. To obtain the negative samples, 20% of these compounds were randomly selected and input into the DUD•E website to generate decoys (Mysinger et al., 2012). In total, 264 decoys are selected randomly as negative samples.

FIGURE 3.

FIGURE 3

The flowchart of hypertension-related active compound identification.

The molecular descriptions of positive and negative compounds were extracted to constitute the hypertension-related dataset. With the collected dataset, our proposed ensemble method was fitted to predict other hypertension-related compounds.

Experiment Results

In this part, the hypertension-related dataset collected is utilized, which contains 88 related compounds and 264 unrelated compounds. AUC, ROC curve, TPR, FRP, Precision, Specificity, and F1 were used to test the performance of our proposed method. In our method, the parameters of nine classifiers were set by default. In FNT, the variable set is defined as T=(c1,c2,,c9) and the operator set is defined as F=(+2,+3,+4,+5) .

Six cross-validation methods were utilized to validate our proposed method. Nine classifiers were also utilized to identify hypertension-related compounds with the same dataset. The ROC curves and AUC performances with the different cross-validation methods are depicted in Figures 49, respectively. From these results, it can be seen that gcForest has the best ROC curves and AUC values among the nine single classifiers. Our proposed ensemble method could perform better than gcForest in terms of ROC and AUC. With 2-cross, 4-cross, 6-cross, 8-cross, 10-cross, and 15-cross validation methods, in terms of AUC, our method is 0.1, 0.3, 0.3, 0.7, 0.3, and 0.4% higher than gcForest, which reveals that our proposed method performs better than nine single classifiers for hypertension-related compound identification.

FIGURE 4.

FIGURE 4

Hypertension-related compound identification performances of ten methods with 2-cross validation methods.

FIGURE 9.

FIGURE 9

Hypertension-related compound identification performances of ten methods with 15-cross validation methods.

FIGURE 5.

FIGURE 5

Hypertension-related compound identification performances of ten methods with 4-cross validation methods.

FIGURE 6.

FIGURE 6

Hypertension-related compound identification performances of ten methods with 6-cross validation methods.

FIGURE 7.

FIGURE 7

Hypertension-related compound identification performances of ten methods with 8-cross validation methods.

FIGURE 8.

FIGURE 8

Hypertension-related compound identification performances of ten methods with 10-cross validation methods.

The TPR, FRP, Precision, Specificity, and F1 performances of the ten methods with the different cross-validation methods are listed in Tables 16, respectively. With 2-cross validation and 4-cross validation methods, LR could obtain the highest TPR performances, which shows that LR could identify more true hypertension-related compounds. For Table 1, RF and SVM have the best FPR performance, which shows that these two methods could identify less non-related compounds as related ones. SVM also has the highest Precision and Specificity performances among the ten methods. For Table 2, RF has the best FPR, Precision, and Specificity performances. Our method performed best in terms of F1, which reveals that it could identify hypertension-related compounds more accurately overall. With 6-cross validation, 8-cross validation, 10-cross validation, and 15-cross validation methods, our methods perform best among ten methods in terms of TPR, FRP, Precision, Specificity, and F1, except that RF has the lowest performance with 4-cross validation methods. The results show that our proposed ensemble method could identify more true hypertension-related and hypertension-unrelated compounds than the other nine single classifiers.

TABLE 1.

Classification performances of ten methods with 2-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.880597 0.019900 0.936508 0.980100 0.907692
gcForest 0.940299 0.054726 0.851351 0.945274 0.893617
AdaBoost 0.791045 0.014925 0.946429 0.985075 0.861789
Decision Tree 0.671642 0.114428 0.661765 0.885572 0.666667
GBDT 0.61194 0.104478 0.66129 0.895522 0.635659
KNN 0.701493 0.039801 0.854545 0.960199 0.770492
LR 0.985075 0.199005 0.622642 0.800995 0.763006
Naive Bayes 0.791045 0.074627 0.779412 0.925373 0.785185
RF 0.671642 0.00995 0.957447 0.99005 0.789474
SVM 0.850746 0.00995 0.966102 0.99005 0.904762

TABLE 6.

Classification performances of ten methods with 15-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.955224 0 1 1 0.977099
gcForest 0.940299 0.0199 0.940299 0.9801 0.940299
AdaBoost 0.880597 0.0199 0.936508 0.9801 0.907692
Decision Tree 0.850746 0.049751 0.850746 0.950249 0.850746
GBDT 0.835821 0.00995 0.965517 0.99005 0.896
KNN 0.895522 0.039801 0.882353 0.960199 0.888889
LR 0.940299 0.034826 0.9 0.965174 0.919708
Naive Bayes 0.955224 0.089552 0.780488 0.910448 0.85906
RF 0.850746 0.00995 0.966102 0.99005 0.904762
SVM 0.880597 0.014925 0.951613 0.985075 0.914729

TABLE 2.

Classification performances of ten methods with 4-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.895522 0.014925 0.952381 0.985075 0.923077
gcForest 0.925373 0.039801 0.885714 0.960199 0.905109
AdaBoost 0.835821 0.0199 0.933333 0.9801 0.88189
Decision Tree 0.686567 0.039801 0.851852 0.960199 0.760331
GBDT 0.671642 0.00995 0.957447 0.99005 0.789474
KNN 0.850746 0.034826 0.890625 0.965174 0.870229
LR 0.940299 0.074627 0.807692 0.925373 0.868966
Naive Bayes 0.80597 0.094527 0.739726 0.905473 0.771429
RF 0.791045 0.00995 0.963636 0.99005 0.868852
SVM 0.776119 0.024876 0.912281 0.975124 0.83871

TABLE 3.

Classification performances of ten methods with 6-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.955224 0.004975 0.984615 0.995025 0.969697
gcForest 0.925373 0.024876 0.925373 0.975124 0.925373
AdaBoost 0.835821 0.0199 0.933333 0.9801 0.88189
Decision Tree 0.656716 0.054726 0.8 0.945274 0.721311
GBDT 0.791045 0.00995 0.963636 0.99005 0.868852
KNN 0.865672 0.049751 0.852941 0.950249 0.859259
LR 0.940299 0.049751 0.863014 0.950249 0.9
Naive Bayes 0.80597 0.094527 0.739726 0.905473 0.771429
RF 0.820896 0.014925 0.948276 0.985075 0.88
SVM 0.791045 0.014925 0.946429 0.985075 0.861789

TABLE 4.

Classification performances of ten methods with 8-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.970149 0.004975 0.984848 0.995025 0.977444
gcForest 0.940299 0.0199 0.940299 0.9801 0.940299
AdaBoost 0.850746 0.014925 0.95 0.985075 0.897638
Decision Tree 0.835821 0.029851 0.903226 0.970149 0.868217
GBDT 0.80597 0.004975 0.981818 0.995025 0.885246
KNN 0.865672 0.044776 0.865672 0.955224 0.865672
LR 0.940299 0.044776 0.875 0.955224 0.906475
Naive Bayes 0.835821 0.089552 0.756757 0.910448 0.794326
RF 0.835821 0.00995 0.965517 0.99005 0.896
SVM 0.791045 0.014925 0.946429 0.985075 0.861789

TABLE 5.

Classification performances of ten methods with 10-cross validation methods.

TPR FRP Precision Specificity F1
Our method 0.955224 0.014925 0.955224 0.985075 0.955224
gcForest 0.925373 0.0199 0.939394 0.9801 0.932331
AdaBoost 0.850746 0.014925 0.95 0.985075 0.897638
Decision Tree 0.850746 0.0199 0.934426 0.9801 0.890625
GBDT 0.776119 0.014925 0.945455 0.985075 0.852459
KNN 0.850746 0.049751 0.850746 0.950249 0.850746
LR 0.940299 0.044776 0.875 0.955224 0.906475
Naive Bayes 0.850746 0.089552 0.76 0.910448 0.802817
RF 0.820896 0.004975 0.982143 0.995025 0.894309
SVM 0.880597 0.014925 0.951613 0.985075 0.914729

Discussion

To investigate the performance of our proposed ensemble further, two classical ensemble methods (averaged ensemble and voting ensemble) were also utilized to infer hypertension-related compounds. The F1 and AUC performances of the hypertension-related compounds by three ensemble methods are depicted in Figure 10 and Figure 11, respectively. From Figures 10, 11, it can be seen that our proposed ensemble method obtained better F1 and AUC performances than averaged and voting ensemble methods, which also shows that our method could identify hypertension-related compounds more accurately than the other two classical ensemble methods.

FIGURE 10.

FIGURE 10

F1 performances of hypertension-related compound by three ensemble methods.

FIGURE 11.

FIGURE 11

AUC performances of hypertension-related compound by three ensemble methods.

Conclusion

To identify hypertension-related closely active compounds, this paper proposed a novel ensemble method based on a flexible neural tree and nine classifiers. In our method, the classification results of nine single classifiers was utilized as the input vector of the flexible neural tree. An FNT model was utilized as a nonlinear ensemble method to identify hypertension-related drug activity. A hybrid evolutionary method based on genetic programming like structure optimization algorithm and simulated annealing is proposed to evolve the FNT model. In order to test the performance of our proposed ensemble method, data were extracted from hypertension-unrelated and hypertension-related compounds collected from up-to-date literature. By the different cross-validation methods, our proposed method obtained better ROC curves and AUC values than nine other single classifiers. Our proposed method also performs better than other single classifiers in terms of TPR, FRP, Precision, Specificity, and F1 in most cases. We also compare our proposed ensemble method with the averaged and voting ensemble methods. The results reveal that our method could identify hypertension-related compounds more accurately than the two classical ensemble methods.

Data Availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author Contributions

WB collected and analyes the data of this work. BY and JW designed the model of this work.

Funding

This work was supported by the talent project of “Qingtan scholar” of Zaozhuang University, Jiangsu Provincial Natural Science Foundation, China (No. SBK2019040953), the Youth Innovation Team of Scientific Research Foundation of the Higher Education Institutions of Shandong Province, China (No. 2019KJM006), the Key Research Program of the Science Foundation of Shandong Province (ZR2020KE001), the Ph.D. research startup foundation of Zaozhuang University (No.2014BS13), and Zaozhuang University Foundation (No. 2015YY02).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Aljumah A., Ahamad M., Siddiqui M. (2011). Predictive Analysis on Hypertension Treatment Using Data Mining Approach in Saudi Arabia. Intell. Inf. Manag. 3 (6), 252–261. 10.4236/iim.2011.36031 [DOI] [Google Scholar]
  2. Amaratunga D., Cabrera J., Sargsyan D., Kostis J. B., Zinonos S., Kostis W. J. (2020). Uses and Opportunities for Machine Learning in Hypertension Research. Int. J. Cardiol. Hypertens. 5, 100027. 10.1016/j.ijchy.2020.100027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Breiman L., Friedman J. H., Olshen R. A., Stone C. J. (1984). Classification and Regression Trees (CART). Biometrics. Monterey, CA: Wadsworth. [Google Scholar]
  4. Breiman L. (2001). Random forest. Machine Learn. 45, 5–32. 10.1023/a:1010933404324 [DOI] [Google Scholar]
  5. Brinks H. L., Eckhart A. D. (2010). Regulation of GPCR Signaling in Hypertension. Biochim. Biophys. Acta (Bba) - Mol. Basis Dis. 1802 (12), 1268–1275. 10.1016/j.bbadis.2010.01.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Cao Y., Miao Q.-G., Liu J.-C., Gao L. (2013). Advance and Prospects of AdaBoost Algorithm. Acta Automatica Sinica 39 (6), 745–758. 10.1016/s1874-1029(13)60052-x [DOI] [Google Scholar]
  7. Chen J., Zhang Y., Wang Y., Jiang P., Zhou G., Li Z., et al. (2021b). Potential Mechanisms of Guizhi Decoction against Hypertension Based on Network Pharmacology and Dahl Salt-Sensitive Rat Model. Chin. Med. 16 (1), 34. 10.1186/s13020-021-00446-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Chen L., Zhu T., Qi J., Zhang Y., Zhang Z., Liu H. (2021a). Pharmacological Mechanism of JiaWeiSiWu Granule in the Treatment of Hypertension Based on Network Pharmacology. Ann. Palliat. Med. 10 (7), 7486–7513. 10.21037/apm-21-1140 [DOI] [PubMed] [Google Scholar]
  9. Chen Y., Abraham A., Bo Y. (2006). Feature Selection and Classification Using Flexible Neural Tree. Neurocomputing 70 (1/3), 305–313. 10.1016/j.neucom.2006.01.022 [DOI] [Google Scholar]
  10. Chen Y., Chen D., Liu S., Yuan F., Guo F., Fang F., et al. (2019). Systematic Elucidation of the Mechanism of Genistein against Pulmonary Hypertension via Network Pharmacology Approach. Int. J. Mol. Sci. 20 (22), 5569. 10.3390/ijms20225569 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Chen Y., Yuan T., Chen D., Liu S., Guo J., Fang L., et al. (2020). Systematic Analysis of Molecular Mechanism of Resveratrol for Treating Pulmonary Hypertension Based on Network Pharmacology Technology. Eur. J. Pharmacol. 888, 173466. 10.1016/j.ejphar.2020.173466 [DOI] [PubMed] [Google Scholar]
  12. Díaz-Uriarte R., Alvarez de Andrés S. (2006). Gene Selection and Classification of Microarray Data Using Random forest. BMC Bioinformatics 7, 3. 10.1186/1471-2105-7-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Essiarab F., Taki H., Malki E. A., Hassar M., Ghalim N., Saile R., et al. (2011). Cardiovascular Risk Factors Prevalence in a Moroccan Population. Eur. J. Scientific Res. 49 (4), 581–589. [Google Scholar]
  14. Fang M., Chen Y., Xue R., Wang H., Chakraborty N., Su T., et al. (2021). A Hybrid Machine Learning Approach for Hypertension Risk Prediction. Neural Comput. Applic. 10.1007/s00521-021-06060-0 [DOI] [Google Scholar]
  15. Furey T. S., Cristianini N., Duffy N., Bednarski D. W., Schummer M., Haussler D. (2000). Support Vector Machine Classification and Validation of Cancer Tissue Samples Using Microarray Expression Data. Bioinformatics 16 (10), 906–914. 10.1093/bioinformatics/16.10.906 [DOI] [PubMed] [Google Scholar]
  16. Hu J., Min J. (2018). Automated Detection of Driver Fatigue Based on EEG Signals Using Gradient Boosting Decision Tree Model. Cogn. Neurodyn 12 (4), 431–440. 10.1007/s11571-018-9485-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Hu Z., Tang J., Wang Z., Zhang K., Zhang L., Sun Q. (2018). Deep Learning for Image-Based Cancer Detection and Diagnosis − A Survey. Pattern Recognition 83, 134–149. 10.1016/j.patcog.2018.05.014 [DOI] [Google Scholar]
  18. Hwang K.-Y., Lee E.-S., Kim G.-W., Hong S.-O., Park J.-S., Kwak M.-S., et al. (2016). Developing Data Quality Management Algorithm for Hypertension Patients Accompanied with Diabetes Mellitus by Data Mining. J. Digital Convergence 14 (7), 309–319. 10.14400/jdc.2016.14.7.309 [DOI] [Google Scholar]
  19. Ji Z., Wang B. (2014). Identifying Potential Clinical Syndromes of Hepatocellular Carcinoma Using PSO-Based Hierarchical Feature Selection Algorithm. Biomed. Res. Int. 2014, 127572. 10.1155/2014/127572 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Ji Z., Wu D., Zhao W., Peng H., Zhao S., Huang D., et al. (2015). Systemic Modeling Myeloma-Osteoclast Interactions under Normoxic/hypoxic Condition Using a Novel Computational Approach. Sci. Rep. 5, 13291. 10.1038/srep13291 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Li X., Guo Y. (2005). Naive Bayesian Classifier Based on Multiple Discriminant Analysis. Inf. Control 34 (5), 580–584. [Google Scholar]
  22. Liang Y., Chen Z., Ward R., Elgendi M. (2018). Photoplethysmography and Deep Learning: Enhancing Hypertension Risk Stratification. Biosensors 8 (4), 101. 10.3390/bios8040101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Liao Y., Vemuri V. R. (2002). Use of K-Nearest Neighbor Classifier for Intrusion Detection. Comput. Security 21 (5), 439–448. 10.1016/s0167-4048(02)00514-x [DOI] [Google Scholar]
  24. Liu T. H., Chen W. H., Chen X. D., Liang Q. E., Tao W. C., Jin Z., et al. (2020). Network Pharmacology Identifies the Mechanisms of Action of TaohongSiwu Decoction Against Essential Hypertension. Med. Sci. Monit. 26, e920682. 10.12659/MSM.920682 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Liu X., Zhang Y., Fu C., Zhang R., Zhou F. (2021). EnRank: An Ensemble Method to Detect Pulmonary Hypertension Biomarkers Based on Feature Selection and Machine Learning Models. Front. Genet. 12, 636429. 10.3389/fgene.2021.636429 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Ma X. L., Zhai X., Liu J. W., Xue X. X., Guo S. Z., Xie H., et al. (2018). Study on the Biological Basis of Hypertension and Syndrome with Liver-Fire Hyperactivity Based on Data Mining Technology. World J. Traditional Chin. Med. 4 (4), 176–180. 10.4103/wjtcm.wjtcm_23_18 [DOI] [Google Scholar]
  27. Maalouf M. (2011). Logistic Regression in Data Analysis: an Overview. Ijdats 3 (3), 281–299. 10.1504/ijdats.2011.041335 [DOI] [Google Scholar]
  28. Morra J. H., Tu Z., Apostolova L. G., Green A. E., Toga A. W., Thompson P. M. (2009). Comparison of AdaBoost and Support Vector Machines for Detecting Alzheimer's Disease Through Automated Hippocampal Segmentation. IEEE Trans. Med. Imaging 29 (1), 30–43. 10.1109/TMI.2009.2021941 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Munshi T., Zuidgeet M., Brussel M., Maarseveen M. V. (2014). Logistic Regression and Cellular Automata-Based Modeling of Retail, Commercial and Residential Development in the City of Ahmedabad India. Cities 38 (2), 88–101. 10.1016/j.cities.2014.02.007 [DOI] [Google Scholar]
  30. Mysinger M. M., Carchia M., Irwin J. J., Shoichet B. K. (2012). Directory of Useful Decoys, Enhanced (DUD-E): Better Ligands and Decoys for Better Benchmarking. J. Med. Chem. 55 (14), 6582–6594. 10.1021/jm300687e [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Owlia M., Bangalore S. (2016). In Hypertensive Patients with Elevated Risk of Cardiovascular Disease, Targeting Systolic Blood Pressure to Less Than 120 Mm Hg Significantly Reduces the Rate of Fatal and Non-fatal Cardiovascular Events as Well as Death from Any Cause. Evid. Based Med. 21 (3), 101. 10.1136/ebmed-2016-110397 [DOI] [PubMed] [Google Scholar]
  32. Ramezankhani A., Kabir A., Pournik O., Azizi F., Hadaegh F. (2016). Classification-based Data Mining for Identification of Risk Patterns Associated with Hypertension in Middle Eastern Population. Medicine 95 (35), e4143. 10.1097/md.0000000000004143 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Rish I. (2001). “An Empirical Study of the Naive Bayes Classifier”, in IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence, Seattle, WA, August 4, 2001 3, 41–46. [Google Scholar]
  34. Sakai K., Sigmund C. D. (2005). Molecular Evidence of Tissue Renin-Angiotensin Systems: A Focus on the Brain. Curr. Sci. Inc 7 (2), 135–140. 10.1007/s11906-005-0088-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Suykens J. A. K., Vandewalle J. (1999). Least Squares Support Vector Machine Classifiers. Neural Process. Lett. 9 (3), 293–300. 10.1023/a:1018628609742 [DOI] [Google Scholar]
  36. Temkin N. R., Holubkov R., Machamer J. E., Winn H. R., Dikmen S. S. (1995). Classification and Regression Trees (CART) for Prediction of Function at 1 Year Following Head Trauma. J. Neurosurg. 82 (5), 764–771. 10.3171/jns.1995.82.5.0764 [DOI] [PubMed] [Google Scholar]
  37. Wang T., He M., Du Y., Chen S., Lv G. (2021). Network Pharmacology Prediction and Pharmacological Verification Mechanism of Yeju Jiangya Decoction on Hypertension. Evid. Based Complement. Alternat Med. 2021, 5579129. 10.1155/2021/5579129 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Wang T., Zhang Y., Lu J., Chai R., Chen X., et al. (2018). Research on the Functional Mechanism of Shengmai Injection Based on Network Pharmacology. J. Pharm. Res. 37 (11), 621–624. 10.13506/j.cnki.jpr.2018.11.001 [DOI] [Google Scholar]
  39. Xu S. N., Li Z., Zhai Y. Y., Yao W. F., Xu J., Liu Q., et al. (2018). Material Basis and Mechanism of Erzhi Pill for Preventing Osteoporosis Based on Network Pharmacology. Chin. Pharm. J. 53 (22), 1913–1920. 10.11669/cpj.2018.22.007 [DOI] [Google Scholar]
  40. Yang B., Chen Y., Jiang M. (2013). Reverse Engineering of Gene Regulatory Networks Using Flexible Neural Tree Models. Neurocomputing 99, 458–466. 10.1016/j.neucom.2012.07.015 [DOI] [Google Scholar]
  41. Yuan F., Chen S. (2011). Model Construction on Efficient Mining Association Rules in Clinical Data of Hypertension. Comput. Eng. Appl. 47 (36), 226–229+233. 10.3778/j.issn.1002-8331.2011.36.062 [DOI] [Google Scholar]
  42. Zhai Z., Tao X., Alami M. M., Shu S., Wang X. (2021). Network Pharmacology and Molecular Docking Combined to Analyze the Molecular and Pharmacological Mechanism of Pinellia Ternata in the Treatment of Hypertension. Cimb 43 (1), 65–78. 10.3390/cimb43010006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Zhang B., Ren J., Cheng Y., Wang B., Wei Z. (2019b). Health Data Driven on Continuous Blood Pressure Prediction Based on Gradient Boosting Decision Tree Algorithm. IEEE Access 7, 32423–32433. 10.1109/access.2019.2902217 [DOI] [Google Scholar]
  44. Zhang Y., Lei L., He J. (2019a). Study on Distribution Rules of TCM Signs and Symptoms and Syndrome Elements in Essential Hypertension Based on Data Mining. Chin. J. Inf. Traditional Chin. Med. 26 (1), 99–104. 10.3969/j.issn.1005-5304.2019.01.023 [DOI] [Google Scholar]
  45. Zhao L., Wu Y. J., Zhang M. Q. (2021). Research Progress of Data Mining in the Treatment of Hypertension by Traditional Chinese Medicine. Food Ther. Health Care 3 (2), 36–46. 10.12032/FTHC20210503 [DOI] [Google Scholar]
  46. Zhou Z. H., Feng J. (2017). Deep Forest: Towards An Alternative to Deep Neural Networks. Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, August 19-25, 2017, 3553–3559. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.


Articles from Frontiers in Genetics are provided here courtesy of Frontiers Media SA

RESOURCES