Skip to main content
Entropy logoLink to Entropy
. 2018 Nov 22;20(12):897. doi: 10.3390/e20120897

Efficient Heuristics for Structure Learning of k-Dependence Bayesian Classifier

Yang Liu 1,2, Limin Wang 1,2, Minghui Sun 1,2,*
PMCID: PMC7512482  PMID: 33266621

Abstract

The rapid growth in data makes the quest for highly scalable learners a popular one. To achieve the trade-off between structure complexity and classification accuracy, the k-dependence Bayesian classifier (KDB) allows to represent different number of interdependencies for different data sizes. In this paper, we proposed two methods to improve the classification performance of KDB. Firstly, we use the minimal-redundancy-maximal-relevance analysis, which sorts the predictive features to identify redundant ones. Then, we propose an improved discriminative model selection to select an optimal sub-model by removing redundant features and arcs in the Bayesian network. Experimental results on 40 UCI datasets demonstrate that these two techniques are complementary and the proposed algorithm achieves competitive classification performance, and less classification time than other state-of-the-art Bayesian network classifiers like tree-augmented naive Bayes and averaged one-dependence estimators.

Keywords: k-dependence Bayesian classifier, minimal-redundancy-maximal-relevance analysis, discriminative model selection

1. Introduction

In machine learning, classification is one of the most important tasks that predicts the unknown class labels according to some known evidence or labeled training samples. Bayesian network classifiers (BNCs) [1,2] provide a classical method to implement classification decision based on the probability framework. In general, BNCs consist of two parts, i.e., B=<G,Θ>. The network structure G is a directed acyclic graph (DAG). Nodes in G represent stochastic variables or features, arc XiXj denotes probabilistic dependency relationships between these two features and Xi is one of immediate parent nodes of Xj, i.e., XiPa(Xj). Parameter Θ quantitatively describes this dependency. Let each instance x be characterized with n values {x1,,xn} for features {X1,,Xn}, and class label c{c1,,cm} is the value of class variable C. Θ contains conditional probability tables θxi|Pa(xi)=pB(xi|Pa(xi)) for each feature.

According to Bayesian theorem [3], BNC makes the classification decision in the following way:

argmaxCp(c|x)=argmaxCp(x,c)p(x)argmaxCp(x,c) (1)

According to the chain rule of joint probability distribution [1], p(x,c) can be calculated as follows:

p(x,c)=p(c)p(x1|c)p(x2|x1,c)p(xn|x1,x2,,xn1,c)=p(c)i=1np(xi|Pa(xi),c). (2)

In this paper, we mainly focus on the restricted BNCs, which suppose that each feature is directly dependent on the class variable C and C does not have any parents. In this paper, we mainly focus on the restricted BNCs, which require that the class variable C be a parent of every feature and no feature be the parent of C.

The k-dependence Bayesian classifier (KDB) is one of the famous restricted BNCs [4]. To achieve the trade-off between structure complexity and classification accuracy, KDB allows to represent different number of interdependencies for different data sizes. During the learning procedure of KDB, it utilizes mutual information between features and class variable to rank and sort all features first. This sorting method gives priority to the features with high relevance between features and class. Feature Xi may be a possible parent feature of Xj if Xi ranks before Xj, not the other way around. Then conditional mutual information between features is used to measure and select significant conditional dependencies. The dependency relationships between features and class, and that between different features, are considered in different learning phases. Obviously, some independent features with high mutual information value may achieve higher rank but demonstrate weak conditional dependencies. To address this issue, Peng et al. [5] propose a first-order incremental feature selection method based on minimal-redundancy-maximal-relevance (mRMR) criterion, which takes into account the maximal relevance between features and class, meanwhile considering the minimal redundancy between features. Its effectiveness has not been proved in the context of KDB.

The structure complexity will increase exponentially as the number of features increases. The features that rank at the end of the order are the least relevant to classification and may be disregarded. Regular KDB does not consider the negative effect caused by redundant features, which may bias the classification results. Many researchers have recognized that using a heuristic wrapper approach to delete redundant features helps minimize zero-one loss on the training samples [6,7,8]. Martínez et al. [9] propose discriminative model selection to select an optimal KDB sub-model which contains feature subset with necessary features. The resulting algorithm not only has the competitive classification performance of generative learning, but also has the excellent expressive power of discriminative learning. At each iteration for model selection, any feature Xi in the order should have k parent features if i>k as KDB defines. However, the dependencies between Xi at the end of the order and the other feature Xj (1jn,ij) may be very weak, and these two features can be assumed to be independent. That is, the dependencies between Xi and Xj may be redundant.

In this paper, we will investigate the feasibility of applying discriminative model selection to remove redundant features and dependencies, and the interoperability of mRMR analysis and discriminative model selection. Section 2 reviews the state-of-the-art restricted BNCs, including naive Bayes (NB), tree-augmented naive Bayes (TAN) and especially KDB. In Section 3 we present the theoretical justification of our proposed algorithm, mRMR-based KDB with discriminative model selection (MMKDB).Section 4 presents a detailed analysis of the experimental results. Finally, we present conclusions in Section 5.

2. Restricted Bayesian Network Classifiers

The classification task in a BNC can be separated into two subtasks, structure learning and parameter learning. The former is to identify the structure of the network, and the latter is to calculate the probability distribution for a given network structure. In the following discussion, we will review some state-of-the-art BNCs from the perspective of structure learning and parameter learning.

NB is the simplest BNC [10,11], since the features are assumed to be conditionally independent given the class variable. The formula of joint probability p(x,c) is presented as follows:

p(x,c)=p(c)i=1np(xi|c). (3)

Note that, for NB, the parameter learning only involves the learning of the probability p(c) and the conditional probability p(xi|c), and the structure learning is not necessary since NB has a definite structure as shown in Figure 1. However, features may be interrelated in practice. Therefore, many researchers have exploited methods to alleviate the conditional independence assumption of NB [12,13,14]. It is worthwhile to mention that, Webb et al. [15] present a new approach, named averaged one-dependence estimators (AODE), to weaken the feature independence assumption by averaging all of the constrained class of classifiers. The class of all such classifiers has all other features depend on a common feature and the class variable.

Figure 1.

Figure 1

The topology structure of NB.

TAN [1] is an extension of NB. It uses a variant of the Chow-Liu algorithm [16] to construct the Bayesian network, and it utilizes conditional mutual information

I(Xi;Xj|C)=xiXixjXjcCp(xi,xj,c)log2p(xi,xj|c)p(xi|c)p(xj|c) (4)

to find a maximum weighted spanning tree. Additional arcs between features are allowed, i.e., dependencies between features can be captured. Each feature in the network has at most one other feature as its parents, except a single feature (the root of the tree), which has only the class variable as its parent. TAN alleviates some of conditional independence assumption of NB and, thus, improves its prediction accuracy at the cost of adding its structure complexity. The joint probability of TAN is calculated by:

p(x,c)=p(c)i=1np(xi|xj,c). (5)

where Xj is the parent of Xi in the tree structure.

KDB is another classical improvement to NB [4]. It allows for most k features to be the parents for each feature. In this sense, NB is a 0-dependence BNC and TAN is a one-dependence BNC. In the real-world domains we find that modeling feature dependencies very often improves classification performance. This is especially true for the KDB, with respect to lower value of k, larger value of k may helps to improve the classification accuracy [4]. Two passes are required for KDB to learn over the training examples. Structure learning is the first pass. Algorithm 1 depicts the structure learning process of KDB. Parameter learning is the second pass. According to the Bayesian network obtained from the former pass, the joint probability of KDB for each instance can be calculated by:

p(x,c)=p(c)i=1np(xi|Pa(xi),c). (6)

where Pa(xi) denotes the parents of Xi in the structure. Suppose there is an ordered feature set {X1,X2,X3,X4}, we give some examples of corresponding structures of KDB classifiers in Figure 2 when given different k values. Corresponding joint probability distributions are shown in Table 1.

Figure 2.

Figure 2

Some examples of corresponding structures of KDB classifiers when given different k values.

Table 1.

Corresponding joint probability distributions of KDB when given different k values.

k Value Joint Probability Distribution
k=1 p(x1,,x4,c)=p(c)p(x1|c)p(x2|x1,c)p(x3|x1,c)p(x4|x1,c)
k=2 p(x1,,x4,c)=p(c)p(x1|c)p(x2|x1,c)p(x3|x1,x2,c)p(x4|x1,x2,c)
k=3 p(x1,,x4,c)=p(c)p(x1|c)p(x2|x1,c)p(x3|x1,x2,c)p(x4|x1,x2,x3,c)
Algorithm 1: Structure learning process of KDB.
graphic file with name entropy-20-00897-i001.jpg

3. The mRMR-Based KDB with the Discriminative Model Selection

To elaborate our motivations for doing selection based on mRMR criterion and discriminative model selection in the context of the KDB classifier, we consider two extreme examples of constructing BNCs over two sets of features. The first feature set contains two perfectly correlated features Xi and Xj, where Xi is an exact copy of Xj. Both Xi and Xj will be included in the network structure of KDB, that is, Xi (or Xj alternatively) will have twice the influence of the other features, which may strongly bias the performance of the classifier. A possible way to improve the classification performance is to eliminate one of the features {Xi,Xj} from the feature set and to construct the classifier over the reduced set of features. The ordered feature set {Xa,Xb,Xc} is the second extreme example and contains non-redundant features to construct a KDB with k=2. Suppose that the values of I(Xc;Xa|C) and I(Xc;Xb|C) are respectively 0.99 and 0.0001. As KDB defines, feature Xc should select Xa and Xb as its parent features in any case. This naturally results in a redundant dependency between Xc and Xb, which may lead to negative effects on the classification performance of KDB and increases the risk of over-fitting at a certain extent.

Therefore, we utilize the sorting method based on an mRMR criterion to identify possible redundant features and discriminative model selection to achieve the aim of removing the redundant features or conditional interdependencies. The usual feature selection based on mutual information in KDB intends to select features that are independent of each other. Instead, the mRMR method tries to select a feature that minimizes the redundancy and maximizes the relevance. As argued by Peng et al. [5], for real data, the features selected in this way will have more or less correlation with each other and the joint effect of these features can lead to very good classification accuracy.

Let S denote the feature set and |S| is the cardinality of S. Given a feature set X={X1,X2,,Xn} and a class variable C, in order to make sure that the selected feature subset is the most appropriate one, two conditions should be met. The first one is the minimum redundancy condition [5]:

MinR(S),whereR(S)=1|S|2Xi,XjSI(Xi;Xj) (7)

where R represents the level of redundancy between features.

And the other one is the maximum relevancy condition [5]:

MaxD(S),whereD(S)=1|S|XiSI(Xi;C) (8)

where D represents the level of relevancy between feature and class variable.

There are two combinations of these two conditions, named MID (Mutual Information Difference) and MIQ (Mutual Information Quotient) [17], which balance the two objectives, maximum relevance and minimum redundancy, in different ways as follows:

MID(S)=max(D(S)R(S)) (9)
MIQ(S)=max(D(S)/R(S)) (10)

As argued by Gulgezen et al. [18], MID produces more stable feature subsets, so in this paper we choose MID as the criterion. Suppose there exists a selected feature subset Sm1, which consists of m1 features, then the m-th feature can be determined by following equation:

MID(Xj)=max{I(Xj;C)1m1XiSm1I(Xi;Xj)} (11)

where XjSSm1.

The feature selection based on mRMR criterion utilizes forward selection strategy, it starts with an empty feature set L and then iteratively add one feature into the L at a time by Equation (11). Sorting all features in this way, we consider the feature subsets {X1,X2,,Xi}, 1in, each feature subset contains i ordered features. That is, for n features there are n alternative feature subsets that could be explored for our proposed algorithm.

From Equation (6) we can observe that the joint probability p(x,c) can be considered as the product of a set of conditional probabilities p(xi|Pa(xi),c). This means that we can build a model space by using a nested method, each model can be built upon the previous one. For an instance x=(x1,x2,,xn), as Table 2 shows, the joint probability p(x,c)2 is obtained by multiplying the conditional probability of feature X2 (i.e., p(x2|Pa(x2),c)) to p(x,c)1 and the joint probability p(x,c)3 is obtained by multiplying the conditional probability of feature X3 (i.e., p(x3|Pa(x3),c)) to p(x,c)2. That is to say, if the model p(x,c)i has been built, it is not necessary to repeat the process of structure learning with feature set {X1,X2,,Xi} for the model p(x,c)i+1. We only need to find parents of the feature Xi+1 in the BN and then multiply the conditional probability p(xi+1|Pa(xi+1),c) with the joint probability p(x,c)i (which has been learnt in the previous model p(x,c)i) to obtain the model p(x,c)i+1. The discriminative model selection framework is derived from the chain rule of BNCs’ joint probability, it firstly constructs a space of sub-models, and then selects the best sub-model by the evaluation function to achieve our purpose in feature selection.

Table 2.

Space of approximate models of KDB with n feature subsets.

Feature Subsets Joint Probability
{X1} p(x,c)1=p(c)p(x1|c)
{X1,X2} p(x,c)2=p(c)p(x1|c)p(x2|Pa(x2),c)
{X1,X2,X3} p(x,c)3=p(c)p(x1|c)p(x2|Pa(x2),c)p(x3|Pa(x3),c)
{X1,X2,X3,,Xn} p(x,c)n=p(c)p(x1|c)p(x2|Pa(x2),c)p(x3|Pa(x3),c)p(xn|Pa(xn),c)

Based on the above observations and discussions, we further improve the framework of discriminative model selection from the view of feature dependencies. To make the idea of the improved framework of discriminative model selection clear in KDB, we restrict that at most two features can be the parents for each feature in the following discussion. As Figure 3 shows, for feature subset {X1,X2,X3}, the corresponding model space of our proposed algorithm MMKDB is composed of BNC30, BNC31 and BNC32. The only difference of these three BNCs is the number of parents for feature X3. We employ the conditional mutual information to assign 0, 1 or 2 features to X3 as parents, respectively. Note that all BNCs with {X1,X2,X3} are built upon BNC21, which is the best BNC for feature subset {X1,X2} and selected by using an evaluation function. Similarly, BNC30, BNC31 and BNC32 also need to be evaluated the classification performance to select the best one. In this way, we can remove not only redundant features but also redundant dependencies between them. That is to say, at each iteration for model selection, any feature Xi should have k parent features, where 0kk if i>k. Note that we employ the root mean squared error (RMSE) [19] as the evaluation function in the procedure of discriminative model selection, which is an effective measure of probability estimates:

RMSE=1txD(1p(c^|x))2 (12)

where D is the training set, t is the number of training examples, c^ is the true class label for the instance x, and p(c^|x) is the estimated posterior probability of the true class given x.

Figure 3.

Figure 3

The examples of corresponding model space of MMKDB for three feature subsets when k=2.

It is worthwhile to note that, in order to avoid over-fitting of sub-models on training examples, we employ the leave-one-out cross-validation (LOOCV) [20] to evaluate the classification performance of each model. Kohavi et al. [21] propose an incremental method to refine the cross-validation. The traditional LOOCV for BNCs recomputes the joint probability of a new model over the training examples for each instance. Differently, the incremental cross-validation firstly calculates the total joint frequency counts for all training examples, and then when testing an instance, temporarily removing its counts from the total counts to calculate the joint probability of corresponding model.

Figure 4 presents the schematic diagram of our proposed algorithm MMKDB. Step 1 sets the order of features by the mRMR sorting method, computes the conditional mutual information between features and class variable through the training examples, and the ordered feature set would be divided into n feature subsets. Steps 2 and 3 correspond to the framework of discriminative model selection. All ordered feature subsets are introduced as the input to construct the corresponding BNCs. Sub-models containing features that rank ahead in the order would be built upon sub-models containing features ranks behind. These n(k+2)2 sub-models form the model space. Each sub-model denotes as BNCsr, where s is the number of feature subsets and r is the number of parents for feature Xs. In the model space, all sub-models are evaluated by using the LOOCV to calculate the values of RMSE through the training examples. According to the chain rule of BNCs’ joint probability, BNCs+1 needs to be built upon BNCs. Thus, only one sub-model that has the lowest value of RMSE would be selected for each feature subset. Finally, there are n alternative local optimum BNCs for n feature subsets. The optimal BNC would be selected from these n sub-models.

Figure 4.

Figure 4

The schematic diagram of MMKDB.

Based on the discussion presented above, we present the pseudo-codes of MMKDB in Algorithm 2. Calculating I(Xi;C) and I(Xi;Xj) respectively need O(tcnv) and O(tn2v2) time, where t is the number of training examples, c is the number of classes, n is the number of features and v is the maximum number of possible values per feature. From Equation (11) we can infer that, if (n>c) the time complexity of step 2 in Algorithm 2 is O(tn2v2), or else O(tcnv). The procedure of computing the conditional mutual information needs O(tcn2v2) time. The space complexity of the table of joint frequencies of all combinations of n features values and the class label is O(cn2v2). Feature ordering needs O(nlogn) time and parent assignment for each feature needs O(n2logn) time. Moreover, classify an instance using the selected sub-model only requires O(cnk) time. That is, the procedure of discriminative model selection needs O(tcnk2) time. So the overall time complexity is O(tcn2v2+tcnk2) for MMKDB and O(tn2v2) for KDB. This is an acceptable result, since k is a user-set parameter. That is, the time complexity of MMKDB scales linearly with the number of training examples, classes and features.

Algorithm 2: Algorithm MMKDB.
graphic file with name entropy-20-00897-i002.jpg

4. Experiments

We run the experiments on a C++ system (GCC 5.4.0) which is specially designed for BNCs. For KDB, with respect to lower value of k, larger value of k may helps to improve the classification accuracy. However, the restrictions of currently available hardware place some requirements on the software. The structure complexity and time complexity will increase exponentionally as k increases. When k=4, due to the amount of memory and CPU available the experimental results of MMKDB on some datasets cannot be achieved. Thus in the following experimental study the maximum value of k is 3.

In our experimental study, we gather a group of datasets from UCI machine learning repository [22]. These datasets are described in Table 3. Missing values are referred to as a distinct value. For each dataset, we discretize quantitative features using 5-bin equal frequency discretization, and we employ the m-estimation (m=1) [23,24] to smooth the probability estimates.

Table 3.

Datasets.

No. Dataset Inst Feature Class No. Dataset Inst Feature Class
1 Contact-Lenses 24 4 3 21 Splice-C4.5 3177 60 3
2 Lung-Cancer 32 56 3 22 Hypo 3772 29 4
3 Labor 57 16 2 23 Sick 3772 29 2
4 Post-Operative 90 8 3 24 Abalone 4177 8 3
5 Zoo 101 16 7 25 Spambase 4601 57 2
6 Promoters 106 57 2 26 Waveform-5000 5000 40 3
7 Echocardiogram 131 6 2 27 Phoneme 5438 7 50
8 Autos 205 25 7 28 Page-Blocks 5473 10 5
9 Audio 226 69 24 29 Mushrooms 8124 22 2
10 Hungarian 294 13 2 30 Thyroid 9169 29 20
11 Dermatology 366 34 6 31 Pendigits 10,992 16 10
12 Horse-Colic 368 21 2 32 Sign 12,546 8 3
13 House-Votes-84 435 16 2 33 Nursery 12,960 8 5
14 Chess 551 39 2 34 Magic 19,020 10 2
15 Crx 690 15 2 35 Letter-Recog 20,000 16 26
16 Vehicle 846 18 4 36 Adult 48,842 14 2
17 Anneal 898 38 6 37 Shuttle 58,000 9 7
18 Led 1000 7 10 38 Connect-4 67,557 42 3
19 Volcanoes 1520 3 4 39 Localization 164,860 5 11
20 Car 1728 6 4 40 Census-Income 299,285 41 2

As a contrast, we also present respectively two extensional version of KDB as follows:

  • KDB with the sorting method based on the mRMR criterion (MKDB).

  • KDB with the discriminative model selection (MSKDB).

Note that, the experiments have been done by using 10 rounds of 10-fold cross-validation, and we employ the zero-one loss to evaluate classification accuracy of different algorithms [25]. Suppose that c is the predicted class label of an algorithm and c^ is the true class label, the value of zero-one loss is calculated as follow:

ξ(c,c^)=1δ(c,c^) (13)

where δ(c,c^)=1 if c=c^ and 0 otherwise.

The detailed zero-one loss results of all alternative algorithms are presented in Table A1 in the Appendix A. In order to give the experimental results an intuitive explanation, we employ the Win/Draw/Loss (W/D/L) records to summarize the number of datasets for different algorithms in the following three situations on a given evaluation function: a win represents an algorithm achieves significant advantages over the other one on a dataset, a loss indicates the opposite case and the draw suggests that these two algorithms perform comparably. Each entry compares the algorithm in the row against the one in the column. We regard a difference as significant between two algorithms if their outcomes of a one-tailed binomial sign test is less than 0.05.

4.1. Impact of Sorting Method Based on the mRMR Criterion and Discriminative Model Selection

In order to investigate the impact of sorting method based on the mRMR criterion, we present the W/D/L records when comparing the zero-one loss results of KDB and MKDB in Table 4. The only difference between KDB and MKDB is the sorting method of features, the former performs the sorting method based on the mutual information and the latter performs the one based on mRMR criterion. From Table 4 we can see that, MKDB achieves significant advantages over KDB and results in W/D/L of 12/25/3. This proves that the sorting method based on the mRMR criterion is superior to the one based on mutual information in KDB. Compared with KDB, there are only three datasets, i.e., Lung-Cancer, House-Votes-84 and Anneal, have higher results of zero-one loss over MKDB, which indicates that MKDB seldom performs worse than KDB, and for many datasets, it substantially improved the classification performance of KDB, such as, the datasets Adult, Dermatology, Labor and Hypo.

Table 4.

W/D/L records when comparing the zero-one loss of MKDB and KDB.

W/D/L KDB
MKDB 12/25/3

In order to explore the effect of discriminative model selection, we present the W/D/L records in terms of zero-one loss between KDB and MSKDB in Table 5. The only difference between these two algorithms is that MSKDB need an extra pass to perform the discriminative model selection through the training examples. As expected, MSKDB achieves lower zero-one loss results more often than KDB, for example, the decrease from 0.1926 to 0.0598 for the dataset Splice-C4.5. Note that MSKDB only performs worse than KDB on one dataset, i.e., Contact-Lenses. We argue that the lack of enough instances is the main reason why MSKDB performs not well on this dataset.

Table 5.

W/D/L records when comparing the zero-one loss of MSKDB and KDB.

W/D/L KDB
MSKDB 8/31/1

4.2. Comparison of MMKDB vs. KDB

According to the zero-one loss, the corresponding comparison with MMKDB and KDB is given in Table 6. Table 6 also presents the W/D/L records of MMKDB over MKDB and MSKDB. As we can see that MMKDB achieves significant advantages than KDB, MKDB and MSKDB, which indicates that the interoperability of mRMR analysis and discriminative model selection is feasible. To further demonstrate the performance of MMKDB over other algorithms, we employ the goal difference (GD) [26]. Suppose there are two classifiers A and B, the value of GD can be computed as follow:

GD(A;B|T)=|win||loss|, (14)

where T is the datasets, |win| and |loss| represent the number of datasets on which A performs better or worse than B, respectively.

Table 6.

W/D/L records when comparing the zero-one loss of KDB, MKDB, MSKDB and MMKDB.

W/D/L KDB MKDB MSKDB
MMKDB 17/21/2 15/23/2 10/30/0

Figure 5 shows the fitting curve of GD(MMKDB; KDB|St) in terms of zero-one loss. The X-axis shows the indexes of different datasets, referred to as t, which correspond to that described in Table 3, and the Y-axis corresponds to the value of GD(MMKDB; KDB|St), where St={Dm|mt} and Dm is the dataset with index m. We categorize datasets according to their size. Datasets with instances ≤1000, >1000 and ≤10,000, >10,000 are represented as small, medium and large size, respectively. Two dotted lines divide the figure into three parts, each part is associated to the corresponding sizes of different datasets. From Figure 5 we can see a clear positive correlation between the values of GD(MMKDB; KDB|St) and the dataset size. As the size of datasets increases, MMKDB achieves significant advantages over KDB on small and medium datasets. When the number of instances >10,000, MMKDB has similar zero-one loss performance to KDB, but it speeds up classification time. Since MMKDB removes not only features but also dependencies between them may be redundant. Thus, we can come to the conclusion that, MMKDB not only retains the privileges of KDB, i.e., the capacity of high dependence representation and the model fitting ability on large datasets, but also improves the model fitting ability on small and medium datasets and enhances the classification efficiency on large datasets. That is, it proves the feasibility of applying discriminative model selection to remove redundant features and dependencies.

Figure 5.

Figure 5

The fitting curve of GD(MMKDB; KDB|St) in terms of zero-one loss.

To further evaluate whether mRMR analysis and discriminative model selection are compatible and the extent to which applying both together improves the classification performance relative to applying each alone, we employ the relative zero-one loss ratio [27]. Given two classifiers A and B, the value of the relative zero-one loss ratio, referred to as RZ(·), is calculated as follow:

RZ(A|B)=1ZAZB (15)

where Z denotes the zero-one loss, and ZA(orB) is the value of zero-one loss of classifier A(orB) on a dataset. The smaller ratio of ZA and ZB, the higher value of RZ(A|B), and the better performance of A.

Figure 6 presents the comparison results of RZ(·) between MKDB, MSKDB, MMKDB and KDB. The X-axis shows the index of dataset, and the Y-axis corresponds to the value of RZ(·). As we can see, on the dataset Audio (No. 9), the values of RZ(MKDB|KDB) and RZ(MSKDB|KDB) are respectively 0.1486 and 0.1328. But when it comes to MMKDB, RZ(MMKDB|KDB) is 0.2184, which is more higher than the former two results. There are also two extreme situations. For the dataset Dermatology (No. 11), RZ(MKDB|KDB) is 0.2006 but RZ(MSKDB|KDB) is 0.0198. That is, MKDB improves significantly with KDB but MSKDB does not. But nevertheless, the value of RZ(MMKDB|KDB) on dataset Dermatology is 0.3482. Another extreme situations is just on the contrary, such as dataset Splice-C4.5 (No. 21), RZ(MKDB|KDB) is 0.0729 and RZ(MSKDB|KDB) is 0.6895, which are very unbalanced. However, RZ(MMKDB|KDB) is 0.6936. That is, the value of RZ(MMKDB|KDB) is always equally well to or better than RZ(MKDB|KDB) and RZ(MSKDB|KDB). Therefore, we can draw a conclusion that mRMR analysis and discriminative model selection are compatible in the framework of KDB.

Figure 6.

Figure 6

The comparison results of relative zero-one loss ratio between MKDB, MSKDB, MMKDB and KDB.

A more intuitive explanation is presented in Figure 7. Each bar represents the mean relative zero-one loss ratio of an algorithm to KDB on 40 datasets. As shown in Figure 7, the values of R¯Z(MKDB|KDB), R¯Z(MSKDB|KDB) and R¯Z(MMKDB|KDB) are respectively 0.0449, 0.0620 and 0.1006. That is, the average improved extent of MMKDB to KDB is obviously higher than MKDB and MSKDB. It proves that the interoperability of mRMR analysis and discriminative model selection is the major reason why applying both together improves the classification performance relative to applying each alone.

Figure 7.

Figure 7

The comparison results of mean relative zero-one loss ratio between MKDB, MSKDB, MMKDB and KDB.

4.3. Comparison of MMKDB vs. NB, TAN and AODE

Table 7 presents the corresponding W/D/L results. As we can see, the zero-one loss results of MMKDB are significantly better when compared to NB and TAN, MMKDB also achieves competitive classification performance over AODE. Figure 8 and Figure 9 respectively present the fitting curves of GD(MMKDB; NB|St) and GD(MMKDB; TAN|St) in terms of zero-one loss. As we can see that, MMKDB has similar performance to NB and TAN on small datasets. However, when the dataset size increased to 1000 instances (Led, No. 18), the prediction performance of MMKDB is obviously better than NB and TAN. That is, MMKDB achieves significant advantages over NB and TAN on medium and large datasets.

Table 7.

W/D/L records in terms of zero-one loss: MMKDB vs. NB, TAN and AODE.

W/D/L NB TAN AODE
MMKDB 26/5/9 24/9/7 20/9/11

Figure 8.

Figure 8

The fitting curve of GD(MMKDB; NB|St) in terms of zero-one loss.

Figure 9.

Figure 9

The fitting curve of GD(MMKDB; NB|St) in terms of zero-one loss.

For one of the famous ensemble BNCs, AODE, Figure 10 presents the corresponding fitting curve of GD(MMKDB; AODE|St) in terms of zero-one loss. we can see that the values of GD decrease when the dataset size ≤1000, which means that MMKDB is very difficult to beat AODE on small datasets. When the dataset >1000 and ≤10,000, MMKDB has similar classification performance to AODE. This makes MMKDB a good substitute to AODE on medium datasets. Note that the fitting curve obviously turns upward when the dataset size >10,000 (the size of dataset Pendigits, No. 31). That is to say, the single BNC, MMKDB, achieves significant advantages over AODE on large datasets.

Figure 10.

Figure 10

The fitting curve of GD(MMKDB; AODE|St) in terms of zero-one loss.

The training and classification time comparisons of NB, TAN, AODE, KDB, MKDB, MSKDB and MMKDB are shown in Figure 11 and Figure 12. Each bar depicts the sum of time on 40 datasets. Although, for most datasets our proposed algorithm MMKDB requires substantially more time for learning than other BNCs, such NB, TAN, AODE and KDB, while the classification time of MMKDB is the least. Note that the training and classification time of MSKDB is similar to MMKDB, and the training and classification time of MKDB is similar to KDB. AODE is an ensemble algorithm. Its classification time increases quadratically with the number of features, and hence much higher for other BNCs in Figure 12. In general, MMKDB saves about 42% of KDB’s classification time and greatly improves the classification performance of KDB at the cost of increasing less training time, and it also enjoys an even greater advantage at classification time compared to NB, TAN and AODE.

Figure 11.

Figure 11

Training time comparisons of NB, TAN, AODE, KDB, MKDB, MSKDB and MMKDB.

Figure 12.

Figure 12

Classification time comparisons of NB, TAN, AODE, KDB, MKDB, MSKDB and MMKDB.

4.4. Global Comparison

In this section, we use the Friedman test for comparison of all alternative algorithms on 40 datasets to perform the significance test [28]. The Friedman test is a non-parametric measure, it can be computed as follows:

FF=(D1)χF2D(g1)χF2 (16)

and

χF2=12Dg(g+1)iRi2g(g+1)24 (17)

where g is the number of alternative algorithms, D is the number of datasets and Ri is the average rank of the i-th algorithm. The best performing algorithm getting the rank of 1, the second best rank 2, . In case of ties, average ranks are assigned. The null hypothesis of the Friedman test is that there is no difference in average ranks. The detailed results of the average rank on 40 datasets are presented in Table A2 in the Appendix A. With 7 algorithms and 40 datasets, the Friedman test is distributed according to the F distribution with g1=71=6 and (g1)(D1)=(71)×(401)=234 degrees of freedom. The critical value of F(6.234) for α=0.05 is 2.1375. The result of Friedman test for zero-one loss, FF=8.6308>2.1375 with p<0.001. Hence, we reject the null-hypothesis. That is to say, the seven algorithms are not equivalent in terms of zero-one loss results.

Figure 13 presents the results of ranking in terms of zero-one loss for all alternative algorithms. The average ranks of different algorithms in terms of zero-one loss on all datasets are respectively {NB(5.34), TAN(4.53), AODE(3.73), KDB(4.20), MKDB(4.00), MSKDB(3.83), MMKDB(2.39)}. That is, the ranking of MMKDB is better than that of other algorithms, followed by AODE, MSKB, MKDB, KDB, TAN and NB.

Figure 13.

Figure 13

The results of ranking in terms of zero-one loss for all alternative algorithms.

In order to further explore which algorithm is significantly different to others, we also perform the Nemenyi test [29] shown in Figure 14. The algorithms are plotted on the dotted line on the basis of their average ranks, which are corresponding to the nodes on the top solid line. Critical Difference (CD) is also shown in the figure. The value of CD is calculated as follow:

CD=qαg(g+1)6D (18)

where the critical value qα for α=0.05 and g=7 is 2.949. For α=0.05 with 7 algorithms and 40 datasets, CD = 2.949×g×(g+1)/(6×D)=2.949×7×(7+1)/(6×40)=1.4245. It is worthwhile to note that, the more leftward the position of algorithms on the black line, the lower the rank will be, and hence the better the performance. The algorithms are connected by a line if their differences are not significant. As the figure shows, NB, TAN, KDB and MKDB have equivalent mean rank. The mean rank of MMKDB is significantly lower than those of NB, TAN, KDB, MKDB and MSKDB. MMKDB also achieves lower mean ranks than AODE, but not significantly so.

Figure 14.

Figure 14

The results of Nemenyi test in terms of zero-one loss for all alternative algorithms.

5. Conclusions

KDB is a famous BNC with the capacity of high dependence representation. To achieve the trade-off between structure complexity and classification accuracy, KDB allows to represent different number of interdependencies for different data sizes. The mRMR analysis and discriminative model selection have both previously been demonstrated to be computationally efficient approaches, the former improves the feature selection method and the latter improves the classification error of KDB. However, on the one hand, the mRMR analysis has not been studied in the context of KDB, and on the other hand, the discriminative model selection still can be improved more, such as removing the redundant dependencies in a BNC. Therefore, in this paper, we investigate the feasibility of applying discriminative model selection to remove redundant features and dependencies, and the interoperability of mRMR analysis and discriminative model selection.

Regular KDB utilizes mutual information between features and class variable to rank and sort all features first. Obviously, some independent features with high mutual information value may achieve higher rank but demonstrate weak conditional dependencies. However, the use of mRMR analysis makes up for this shortcoming. Moreover, KDB does not consider the negative effect caused by redundant features, which may bias the classification results. We use the discriminative model selection to achieve the aim of removing the redundant features and arcs in the Bayesian network.

We conduct experiments on 40 UCI datasets to explore the impact of a sorting method based on the mRMR criterion and discriminative model selection. The advantages of MKDB and MSKDB over KDB in terms of zero-one loss, respectively, demonstrate that each technique can help reduce KDB’s classification error. The advantages of MMKDB over KDB, MKDB and MSKDB further demonstrate that the interoperability of these two techniques is feasible. That is, there is strong synergy between the mRMR analysis and discriminative model selection in KDB and they can operate in tandem to reduce the classification error of KDB more effectively than does either in isolation. The fitting curve of goal difference between MMKDB and KDB clarifies the superior performance of the MMKDB on datasets of different scales. MMKDB not only retains the privileges of KDB, i.e., the capacity of high dependence representation and the model fitting ability on large datasets, but also improves the model fitting ability on small and medium datasets and enhances the classification efficiency on large datasets. These two techniques help save about 42% of KDB’s classification time and greatly improve the classification performance of KDB. Besides, we also have compared MMKDB against other state-of-the-art BNCs, such as NB, TAN and AODE. The results demonstrate that MMKDB achieves significant advantages over NB and TAN on medium and large datasets, and over AODE on large datasets in terms of classification performance. We additionally conduct a set of focused tests for some significance analysis, such as the Friedman test and the Nemenyi test. The results showed that the mean rank of MMKDB is significantly lower than those of NB, TAN, KDB, MKDB and MSKDB. MMKDB also achieves lower mean rank than AODE, but not significantly so.

Acknowledgments

This work was supported by the National Science Foundation of China (Grant No. 61272209 and No. 61872164).

Appendix A

This appendix presents the detailed zero-one loss results of NB, TAN, AODE, KDB, MKDB, MSKDB and MMKDB in Table A1, and the ranks of different algorithms on 40 datasets are shown in Table A2.

Table A1.

Detailed zero-one loss results of NB, TAN, AODE, KDB, MKDB, MSKDB and MMKDB.

Dataset NB TAN AODE KDB MKDB MSKDB MMKDB
Contact-Lenses 0.3788 0.3788 0.3788 0.2946 0.2888 0.3713 0.3257
Lung-Cancer 0.4419 0.5997 0.5050 0.5050 0.6806 0.5259 0.5496
Labor 0.0355 0.0531 0.0531 0.0709 0.0347 0.0521 0.0343
Post-Operative 0.3478 0.3704 0.3366 0.3928 0.3758 0.3740 0.3692
Zoo 0.0300 0.0100 0.0300 0.0600 0.0490 0.0490 0.0387
Promoters 0.0763 0.1334 0.1334 0.3240 0.2896 0.3362 0.2582
Echocardiogram 0.3393 0.3315 0.3238 0.3315 0.3325 0.3325 0.3282
Autos 0.3153 0.2167 0.2069 0.2020 0.1931 0.1931 0.1859
Audio 0.2413 0.2949 0.2055 0.3486 0.2968 0.3022 0.2724
Hungarian 0.1615 0.1718 0.1684 0.2027 0.1953 0.2088 0.1928
Dermatology 0.0193 0.0331 0.0166 0.0690 0.0541 0.0676 0.0453
Horse-Colic 0.2196 0.2113 0.2031 0.2882 0.2475 0.1614 0.1620
House-Votes-84 0.0952 0.0558 0.0534 0.0488 0.0524 0.0478 0.0494
Chess 0.1136 0.0935 0.1008 0.0733 0.0737 0.0701 0.0727
Crx 0.1391 0.1493 0.1361 0.1639 0.1549 0.1478 0.1431
Vehicle 0.3963 0.2972 0.2925 0.2901 0.2809 0.2937 0.2784
Anneal 0.0383 0.0112 0.0090 0.0090 0.0097 0.0088 0.0087
Led 0.2697 0.2687 0.2707 0.2687 0.2633 0.2633 0.2599
Volcanoes 0.3349 0.3349 0.3349 0.3349 0.3283 0.3283 0.3240
Car 0.1414 0.0573 0.0824 0.0205 0.0201 0.0201 0.0198
Splice-C4.5 0.0448 0.0471 0.0369 0.1926 0.1786 0.0598 0.0590
Hypo 0.0139 0.0142 0.0096 0.0139 0.0118 0.0097 0.0091
Sick 0.0311 0.0260 0.0276 0.0219 0.0218 0.0213 0.0202
Abalone 0.4810 0.4633 0.4517 0.4647 0.4380 0.4555 0.4423
Spambase 0.1025 0.0676 0.0679 0.0663 0.0671 0.0652 0.0657
Waveform-5000 0.2026 0.1862 0.1477 0.2699 0.2584 0.2600 0.2503
Phoneme 0.2641 0.2760 0.2416 0.2075 0.2033 0.1890 0.1865
Page-Blocks 0.0625 0.0419 0.0341 0.0332 0.0341 0.0317 0.0313
Mushrooms 0.0198 0.0001 0.0001 0.0000 0.0000 0.0000 0.0000
Thyroid 0.1122 0.0727 0.0708 0.0726 0.0714 0.0707 0.0697
Pendigits 0.1193 0.0324 0.0202 0.0444 0.0444 0.0436 0.0438
Sign 0.3622 0.2783 0.2849 0.2311 0.2265 0.2265 0.2236
Nursery 0.0983 0.0661 0.0737 0.0182 0.0175 0.0178 0.0176
Magic 0.2261 0.1692 0.1770 0.1661 0.1618 0.1629 0.1597
Letter-Recog 0.2550 0.1313 0.0892 0.1028 0.0960 0.1008 0.0981
Adult 0.1608 0.1394 0.1508 0.1404 0.1183 0.1355 0.1343
Shuttle 0.0039 0.0015 0.0008 0.0008 0.0008 0.0008 0.0007
Connect-4 0.2811 0.2378 0.2444 0.2166 0.2197 0.2124 0.2168
Localization 0.5005 0.3611 0.3632 0.3097 0.2974 0.3035 0.2996
Census-Income 0.2387 0.0634 0.1014 0.0509 0.0506 0.0488 0.0483

Table A2.

Ranks of different algorithms on 40 datasets.

Dataset NB TAN AODE KDB MKDB MSKDB MMKDB
Contact-Lenses 5.5 5.5 5.5 1.5 1.5 5.5 3.0
Lung-Cancer 1.0 6.0 2.5 2.5 7.0 4.0 5.0
Labor 2.5 5.0 5.0 7.0 2.5 5.0 1.0
Post-Operative 2.0 3.0 1.0 7.0 6.0 5.0 4.0
Zoo 2.5 1.0 2.5 7.0 5.5 5.5 4.0
Promoters 1.0 2.5 2.5 6.0 5.0 7.0 4.0
Echocardiogram 6.0 2.5 1.0 2.5 6.0 6.0 4.0
Autos 7.0 6.0 5.0 4.0 2.5 2.5 1.0
Audio 2.0 4.0 1.0 7.0 5.0 6.0 3.0
Hungarian 1.0 3.0 2.0 6.0 5.0 7.0 4.0
Dermatology 2.0 3.0 1.0 6.5 5.0 6.5 4.0
Horse-Colic 5.0 4.0 3.0 7.0 6.0 1.0 2.0
House-Votes-84 7.0 6.0 4.5 1.5 4.5 1.5 3.0
Chess 7.0 5.0 6.0 2.0 4.0 1.0 3.0
Crx 2.0 4.0 1.0 7.0 6.0 5.0 3.0
Vehicle 7.0 5.0 4.0 3.0 2.0 6.0 1.0
Anneal 7.0 6.0 3.0 3.0 5.0 3.0 1.0
Led 6.0 3.5 7.0 3.5 3.5 3.5 1.0
Volcanoes 4.5 4.5 4.5 4.5 4.5 4.5 1.0
Car 7.0 5.0 6.0 3.0 3.0 3.0 1.0
Splice-C4.5 2.0 3.0 1.0 7.0 6.0 5.0 4.0
Hypo 5.5 7.0 2.0 5.5 4.0 3.0 1.0
Sick 7.0 5.0 6.0 3.0 4.0 2.0 1.0
Abalone 7.0 4.0 3.0 5.5 1.0 5.5 2.0
Spambase 7.0 4.0 5.0 1.0 6.0 2.0 3.0
Waveform-5000 3.0 2.0 1.0 7.0 5.0 6.0 4.0
Phoneme 6.0 7.0 5.0 3.5 3.5 2.0 1.0
Page-Blocks 7.0 6.0 4.0 3.0 5.0 2.0 1.0
Mushrooms 7.0 5.5 5.5 2.5 2.5 2.5 2.5
Thyroid 7.0 5.0 1.0 4.0 6.0 3.0 2.0
Pendigits 7.0 2.0 1.0 3.5 6.0 3.5 5.0
Sign 7.0 5.0 6.0 3.0 3.0 3.0 1.0
Nursery 7.0 5.0 6.0 3.5 1.0 3.5 2.0
Magic 7.0 5.0 6.0 3.5 2.0 3.5 1.0
Letter-Recog 7.0 6.0 1.0 4.5 2.0 4.5 3.0
Adult 7.0 4.0 6.0 5.0 1.0 3.0 2.0
Shuttle 7.0 6.0 3.5 3.5 3.5 3.5 1.0
Connect-4 7.0 5.0 6.0 1.5 4.0 1.5 3.0
Localization 7.0 5.0 6.0 3.5 1.0 3.5 2.0
Census-Income 7.0 5.0 6.0 3.0 4.0 2.0 1.0
Mean rank 5.34 4.53 3.73 4.20 4.00 3.83 2.39

Author Contributions

All authors have contributed to the study and preparation of the article. The 1st author conceived the idea, derived equations and wrote the paper. The 2nd author and the 3rd author did the analysis and finished the programming work. All authors have read and approved the final manuscript.

Funding

This work was supported by the National Science Foundation of China (Grant No. 61272209 and No. 61872164).

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Friedman N., Geiger D., Goldszmidt M. Bayesian network classifiers. Mach. Learn. 1997;29:131–163. doi: 10.1023/A:1007465528199. [DOI] [Google Scholar]
  • 2.Bielza C., Larrañaga P. Discrete Bayesian network classifiers: A survey. ACM Comput. Surv. 2014;47 doi: 10.1145/2576868. [DOI] [Google Scholar]
  • 3.Pearl J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann; Burlington, MA, USA: 1988. [Google Scholar]
  • 4.Sahami M. Learning Limited Dependence Bayesian Classifiers; Proceedings of the Second International Conference on Knowledge Discovery and Data Mining; Portland, OR, USA. 2–4 August 1996; pp. 335–338. [Google Scholar]
  • 5.Peng H., Long F., Ding C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. 2005;27:1226–1238. doi: 10.1109/TPAMI.2005.159. [DOI] [PubMed] [Google Scholar]
  • 6.Bermejo P., Gámez J.A., Puerta J.M. Speeding up incremental wrapper feature subset selection with Naive Bayes classifier. Knowl.-Based Syst. 2014;55:140–147. doi: 10.1016/j.knosys.2013.10.016. [DOI] [Google Scholar]
  • 7.Zare H., Niazi M. Relevant based structure learning for feature selection. Eng. Appl. Artif. Intell. 2016;55:93–102. doi: 10.1016/j.engappai.2016.06.001. [DOI] [Google Scholar]
  • 8.Nhaila H., Elmaizi A., Sarhrouni E., Hammouch A. New wrapper method based on normalized mutual information for dimension reduction and classification of hyperspectral images; Proceedings of the IEEE Fourth International Conference on Optimization and Applications (ICOA); Mohammedia, Morocco. 26–27 April 2018; pp. 1–7. [Google Scholar]
  • 9.Martínez A.M., Webb G.I., Chen S., Zaidi N.A. Scalable learning of Bayesian network classifiers. J. Mach. Learn. Res. 2016;17:1–35. [Google Scholar]
  • 10.Duda R.O., Hart P.E. Pattern Classification and Scene Analysis. Wiley; New York, NY, USA: 1973. A Wiley-Interscience Publication. [Google Scholar]
  • 11.Lewis D.D. Naive (Bayes) at forty: The independence assumption in information retrieval; Proceedings of the European Conference on Machine Learning; Chemnitz, Germany. 21–24 April 1998; Berlin/Heidelberg, Germany: Springer; 1998. pp. 4–15. [Google Scholar]
  • 12.Frank E., Hall M., Pfahringer B. Locally weighted naive bayes; Proceedings of the Nineteenth Conference on Uncertainty in Artificial Intelligence; Edmonton, AB, Canada. 1–4 August 2002; Burlington, MA, USA: Morgan Kaufmann Publishers Inc.; 2002. pp. 249–256. [Google Scholar]
  • 13.Jiang L., Zhang H., Cai Z. A novel Bayes model: Hidden naive Bayes. IEEE Trans. Knowl. Data Eng. 2009;21:1361–1371. doi: 10.1109/TKDE.2008.234. [DOI] [Google Scholar]
  • 14.Langley P., Sage S. Induction of selective Bayesian classifiers; Proceedings of the 10th International Conference Uncertainty Artificial Intelligence; Washington, DC, USA. 29–31 July 1994; pp. 399–406. [Google Scholar]
  • 15.Webb G.I., Boughton J.R., Wang Z. Not so naive Bayes: Aggregating one-dependence estimators. Mach. Learn. 2005;58:5–24. doi: 10.1007/s10994-005-4258-6. [DOI] [Google Scholar]
  • 16.Chow C., Liu C. Approximating discrete probability distributions with dependence trees. IEEE Trans. Inform. Theory. 1968;14:462–467. doi: 10.1109/TIT.1968.1054142. [DOI] [Google Scholar]
  • 17.Ding C., Peng H. Minimum redundancy feature selection from microarray gene expression data. J. Bioinform. Comput. Biol. 2005;3:185–205. doi: 10.1142/S0219720005001004. [DOI] [PubMed] [Google Scholar]
  • 18.Gulgezen G., Cataltepe Z., Yu L. Stable and accurate feature selection; Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Bled, Slovenia. 7–11 September 2009; Berlin/Heidelberg, Germany: Springer; 2009. pp. 455–468. [Google Scholar]
  • 19.Hyndman R.J., Koehler A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006;22:679–688. doi: 10.1016/j.ijforecast.2006.03.001. [DOI] [Google Scholar]
  • 20.Hastie T., Tibshirani R., Friedman J. The Elements of Statistical Learning. Springer; New York, NY, USA: 2009. Unsupervised learning; pp. 485–585. [Google Scholar]
  • 21.Kohavi R. The power of decision tables; Proceedings of the European Conference on Machine Learning; Crete, Greece. 25–27 April 1995; Berlin/Heidelberg, Germany: Springer; 1995. pp. 174–189. [Google Scholar]
  • 22.Bache K., Lichman M. UCI Machine Learning Repository. [(accessed on 1 June 2018)]; Available online: https://archive.ics.uci.edu/ml/datasets.html.
  • 23.Cestnik B. Estimating probabilities: A crucial task in machine learning; Proceedings of the 9th European Conference on Artificial Intelligence; Stockholm, Sweden. 6–10 August 1990; pp. 147–149. [Google Scholar]
  • 24.Zaidi N.A., Cerquides J., Carman M.J., Webb G.I. Alleviating naive bayes attribute independence assumption by attribute weighting. J. Mach. Learn. Res. 2013;14:1947–1988. [Google Scholar]
  • 25.Kohavi R., Wolpert D. Bias Plus Variance Decomposition for Zero-One Loss Functions; Proceedings of the Thirteenth International Conference on Machine Learning; Bari, Italy. 3–6 July 1996; pp. 275–283. [Google Scholar]
  • 26.Duan Z., Wang L. K-Dependence Bayesian Classifier Ensemble. Entropy. 2017;19:651. doi: 10.3390/e19120651. [DOI] [Google Scholar]
  • 27.Wang L.M., Zhao H.Y., Sun M.H., Ning Y. General and Local: Averaged k-Dependence Bayesian Classifiers. Entropy. 2015;17:4134–4154. doi: 10.3390/e17064134. [DOI] [Google Scholar]
  • 28.Demšar J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006;7:1–30. [Google Scholar]
  • 29.Nemenyi P. Ph.D. Thesis. Princeton University; Princeton, NJ, USA: 1963. Distribution-Free Multiple Comparisons. [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES