Abstract
The ease of interpretation of a classification model is essential for the task of validating it. Sometimes it is required to clearly explain the classification process of a model’s predictions. Models which are inherently easier to interpret can be effortlessly related to the context of the problem, and their predictions can be, if necessary, ethically and legally evaluated. In this paper, we propose a novel method to generate rule-based classifiers from categorical data that can be readily interpreted. Classifiers are generated using a multi-objective optimization approach focusing on two main objectives: maximizing the performance of the learned classifier and minimizing its number of rules. The multi-objective evolutionary algorithms ENORA and NSGA-II have been adapted to optimize the performance of the classifier based on three different machine learning metrics: accuracy, area under the ROC curve, and root mean square error. We have extensively compared the generated classifiers using our proposed method with classifiers generated using classical methods such as PART, JRip, OneR and ZeroR. The experiments have been conducted in full training mode, in 10-fold cross-validation mode, and in train/test splitting mode. To make results reproducible, we have used the well-known and publicly available datasets Breast Cancer, Monk’s Problem 2, Tic-Tac-Toe-Endgame, Car, kr-vs-kp and Nursery. After performing an exhaustive statistical test on our results, we conclude that the proposed method is able to generate highly accurate and easy to interpret classification models.
Keywords: multi-objective evolutionary algorithms, rule-based classifiers, interpretable machine learning, categorical data
1. Introduction
Supervised Learning is the branch of Machine Learning (ML) [1] focused on modeling the behavior of systems that can be found in the environment. Supervised models are created from a set of past records, each one of which, usually, consists of an input vector labeled with an output. A supervised model is an algorithm that simulates the function that maps inputs with outputs [2]. The best models are those that predict the output of new inputs in the most accurate way. Thanks to modern computing capabilities, and to the digitization of ever-increasing quantities of data, nowadays, supervised learning techniques play a leading role in many applications. The first classification systems date back to the 1990s; in those days, researchers were focused on both precision and interpretability, and the systems to be modeled were relatively simple. Years later, when it became necessary to model more difficult behaviors, the researchers focused on developing more and more precise models, leaving aside the interpretability. Artificial Neural Networks (ANN) [3], and, more recently, Deep Learning Neural Networks (DLNN) [4], as well as Support Vector Machines (SVM) [5], and Instance-based Learning (IBL) [6] are archetypical examples of this approach. A DLNN, for example, is a large mesh of ordered nodes arranged in a hierarchical manner and composed of a huge number of variables. DLNNs are capable of modeling very complex behaviors, but it is extremely difficult to understand the logic behind their predictions, and similar considerations can be drawn for SVNs and IBLs, although the underlying principles are different. These models are known as black-box methods. While there are applications in which knowing the ratio behind a prediction is not necessarily relevant, (e.g., predicting a currency’s future value, whether or not a user clicks on an advert or the amount of rain in a certain area), there are other situations where the interpretability of a model plays a key role.
The interpretability of classification systems refers to the ability they have to explain their behavior in a way that is easily understandable by a user [7]. In other words, a model is considered interpretable when a human is able to understand the logic behind its prediction. In this way, Interpretable classification models allow external validation by an expert. Additionally, there are certain disciplines such as medicine, where it is essential to provide information about decision making for ethical and human reasons. Likewise, when a public institution asks an authority for permission to investigate an alleged offender, or when the CEO of a certain company wants to take a difficult decision which can seriously change the direction of the company, some kind of explanations to justify these decisions may be required. In these situations, using transparent (also called grey-box) models is recommended. While there is a general consensus on how the performance of a classification system is measured (popular metrics include accuracy, area under the ROC curve, and root mean square error), there is no universally accepted metric to measure the interpretability of the models. Nor is there an ideal balance between the interpretability and performance of classification systems but this depends on the specific application domain. However, the rule of thumb says that the simpler a classification system is, the easier it is to interpret. Rule-based Classifiers (RBC) [8,9] are among the most popular interpretable models, and some authors define the degree of interpretability of an RBC as the number of its rules or as the number of axioms that the rules have. These metrics tend to reward models with fewer rules as simple as possible [10,11]. In general, RBCs are classification learning systems that achieve a high level of interpretability because they are based on a human-like logic. Rules follow a very simple schema:
IF (Condition 1) and (Condition 2) and … (Condition N) THEN (Statement) |
and the fewer rules the models have and the fewer conditions and attributes the rules have, the easier it will be for a human to understand the logic behind each classification. In fact, RBCs are so natural in some applications that they are used to interpret other classification models such as Decision Trees (DT) [12]. RBCs constitute the basis of more complex classification systems based on fuzzy logic [13] such as LogitBoost or AdaBoost [14].
Our approach investigates the conflict between accuracy and interpretability as a multi-objective optimization problem. We define a solution as a set of rules (that is, a classifier), and establish two objectives to be maximized: interpretability and accuracy. We decided to solve this problem by applying multi-objective evolutionary algorithms (MOEA) [15,16] as meta-heuristics, and, in particular, two known algorithms: NSGA-II [15] and ENORA [17]. They are both state-of-the-art evolutionary algorithms which have been applied, and compared, on several occasions [18,19,20]. NSGA-II is very well-known and has the advantage of being available in many implementations, while ENORA generally has a higher performance. In the current literature, MOEAs are mainly used for learning RBCs based on fuzzy logic [18,21,22,23,24,25,26]. However, Fuzzy RBCs are designed for numerical data, from which fuzzy sets are constructed and represented by linguistic labels. In this paper, on the contrary, we are interested in RBCs for categorical data, for which a novel approach is necessary.
This paper is organized as follows. In Section 2, we introduce multi-objective constrained optimization, the evolutionary algorithms ENORA and NSGA-II, and the well-known rule-based classifier learning systems PART, JRip, OneR and ZeroR. In Section 3, we describe the structure of an RBC for categorical data, and we propose the use of multi-objective optimization for the task of learning a classifier. In Section 4, we show the result of our experiments, performed on the well-known publicly accessible datasets Breast Cancer, Monk’s Problem 2, Tic-Tac-Toe-Endgame, Car, kr-vs-kp and Nursery. The experiments allow a comparison among the performance of the classifiers learned by our technique against those of classifiers learned by PART, JRip, OneR and ZeroR, as well as a comparison between ENORA and NSGA-II for the purposes of this task. In Section 5, the results are analyzed and discussed, before concluding in Section 6. Appendix A and Appendix B show the tables of the statistical tests results. Appendix C shows the symbols and the nomenclature used in the paper.
2. Background
2.1. Multi-Objective Constrained Optimization
The term optimization [27] refers to the selection of the best element, with regard to some criteria, from a set of alternative elements. Mathematical programming [28] deals with the theory, algorithms, methods and techniques to represent and solve optimization problems. In this paper, we are interested in a class of mathematical programming problems called multi-objective constrained optimization problems [29], which can be formally defined, for l objectives and m constraints, as follows:
(1) |
where (usually called objectives) and are arbitrary functions. Optimization problems can be naturally separated into two categories: those with discrete variables, which we call combinatorial, and those with continuous variables. In combinatorial problems, we are looking for objects from a finite, or countably infinite, set , where objects are typically integers, sets, permutations, or graphs. In problems with continuous variables, instead, we look for real parameters belonging to some continuous domain. In Equation (1), represents the set of decision variables, where is the domain for each variable , .
Now, let be the set of all feasible solutions to Equation (1). We want to find a subset of solutions called non-dominated set (or Pareto optimal set). A solution is non-dominated if there is no other solution that dominates , and a solution dominates if and only if there exists i () such that improves , and for every i (), does not improve . In other words, dominates if and only if is better than for at least one objective, and not worse than for any other objective. The set of non dominated solutions of Equation (1) can be formally defined as:
where:
Once the set of optimal solutions is available, the most satisfactory one can be chosen by applying a preference criterion. When all the functions are linear, then the problem is a linear programming problem [30], which is the classical mathematical programming problem and for which extremely efficient algorithms to obtain the optimal solution exist (e.g., the simplex method [31]). When any of the functions is non-linear then we have a non-linear programming problem [32]. A non-linear programming problem in which the objectives are arbitrary functions is, in general, intractable. In principle, any search algorithm can be used to solve combinatorial optimization problems, although it is not guaranteed that they will find an optimal solution. Metaheuristics methods such as evolutionary algorithms [33] are typically used to find approximate solutions for complex multi-objective optimization problems, including feature selection and fuzzy classification.
2.2. The Multi-Objective Evolutionary Algorithms ENORA and NSGA-II
The MOEA ENORA [17] and NSGA-II [15] use a strategy (Algorithm 1) with , where corresponds to the number of parents and refers to the number of children ( is the population size), with binary tournament selection (Algorithm 2) and a rank function based on Pareto fronts and crowding (Algorithms 3 and 4). The difference between NSGA-II and ENORA is how the calculation of the ranking of the individuals in the population is performed. In ENORA, each individual belongs to a slot (as established in [34]) of the objective search space, and the rank of an individual in a population is the non-domination level of the individual in its slot. On the other hand, in NSGA-II, the rank of an individual in a population is the non-domination level of the individual in the whole population. Both ENORA and NSGA-II MOEAs use the same non-dominated sorting algorithm, the fast non-dominated sorting [35]. It compares each solution with the rest of the solutions and stores the results so as to avoid duplicate comparisons between every pair of solutions. For a problem with l objectives and a population with N solutions, this method needs to conduct objective comparisons, which means that it has a time complexity of [36]. However, ENORA distributes the population in N slots (in the best case), therefore, the time complexity of ENORA is in the worst case and in the best case.
Algorithm 1 strategy for multi-objective optimization. |
Require: {Number of generations} Require: {Number of individuals in the population}
|
Algorithm 2 Binary tournament selection. |
Require:P {Population}
|
Algorithm 3 Rank-crowding function. |
Require:P {Population} Require: {Individuals to compare}
|
The main reason ENORA and NSGA-II behave differently is as follows. NSGA-II never selects the individual dominated by the other in the binary tournament, while, in ENORA, the individual dominated by the other may be the winner of the tournament. Figure 1 shows this behavior graphically. For example, if individuals B and C are selected for a binary tournament with NSGA-II, individual B beats C because B dominates C. Conversely, individual C beats B with ENORA because individual C has a better rank in his slot than individual B. In this way, ENORA allows the individuals in each slot to evolve towards the Pareto front encouraging diversity. Even though in ENORA the individuals of each slot may not be the best of the total individuals, this approach generates a better hypervolume than that of NSGA-II throughout the evolution process.
ENORA is our MOEA, on which we are intensively working over the last decade. We have applied ENORA to constrained real-parameter optimization [17], fuzzy optimization [37], fuzzy classification [18], feature selection for classification [19] and feature selection for regression [34]. In this paper, we apply it to rule-based classification. NSGA-II algorithm was designed by Deb et al. and has been proved to be a very powerful and fast algorithm in multi-objective optimization contexts of all kinds. Most researchers in multi-objective evolutionary computation use NSGA-II as a baseline to compare the performance of their own algorithms. Although NSGA-II was developed in 2002 and remains a state-of-the-art algorithm, it is still a challenge to improve on it. There is a recently updated improved version for many-objective optimization problems called NSGA-III [38].
Algorithm 4 Crowding_distance function. |
Require:P {Population} Require: P {Population} Require: I {Individual} Require: l {Number of objectives}
|
2.3. PART
PART (Partial DT Method [39]) is a widely used rule learning algorithm that was developed at the University of Waikato in New Zealand [40]. Experiments show that it is a very efficient algorithm in terms of both computational performance and results. PART combines the divide-and-conquer strategy typical of decision tree learning with the separate-and-conquer strategy [41] typical of rule learning, as follows. A decision tree is first constructed (using C4.5 algorithm [42]), and the leaf with the highest coverage is converted into a rule. Then, the set of instances that are covered by that rule are discarded, and the process starts over. The result is an ordered set of rules, completed by a default rule that applies to instances that do not meet any previous rule.
2.4. JRip
JRip is a fast and optimized implementation in Weka of the famous RIPPER (Repeated Incremental Pruning to Produce Error Reduction) algorithm [43]. RIPPER was proposed in [44] as a more efficient version of the incrementally reduced error pruning (IREP) rule learner developed in [45]. IREP and RIPPER work in a similar manner. They begin with a default rule and, using a training dataset, attempt to learn rules that predict exceptions to the default. Each rule learned is a conjunction of propositional literals. Each literal corresponds to a split of the data based on the value of a single feature. This family of algorithms, similar to decision trees, has the advantage of being easy to interpret, and experiments show that JRip is particularly efficient in large datasets. RIPPER and IREP use a strategy based on the separate-and-conquer method to generate an ordered set of rules that are extracted directly from the dataset. The classes are examined one by one, prioritizing those that have more elements. These algorithms are based on four basic steps (growing, pruning, optimizing and selecting) applied repetitively to each class until a stopping condition is met [44]. These steps can be summarized as follows. In the growing phase, rules are created taking into account an increasing number of predictors until the stopping criterion is satisfied (in the Weka implementation, the procedure selects the condition with the highest information gain). In the pruning phase redundancy is eliminated and long rules are reduced. In the optimization phase, the rules generated in the previous steps are improved (if possible) by adding new attributes or by adding new rules. Finally, in the selection phase, the best rules are selected and the others discarded.
2.5. OneR
OneR (One Rule) is a very simple, while reasonably accurate, classifier based on a frequency table. First, OneR generates a set of rules for each attribute of the dataset, and, then, it selects only one rule from that set—the one with the lowest error rate [46]. The set of rules is created using a frequency table constructed for each predictor of the class, and numerical classes are converted into categorical values.
2.6. ZeroR
Finally, ZeroR (Zero Rules [40]) is a classifier learner that does not create any rules and uses no attributes. ZeroR simply creates the class classification table by selecting the most frequent value. Such a classifier is obviously the simplest possible one, and its capabilities are limited to the prediction of the majority class. In the literature, it is not used for practical classifications tasks, but as a generic reference to measure the performance of other classifiers.
3. Multi-Objective Optimization for Categorical Rule-Based Classification
In this section, we propose a general schema for an RBC specifically designed for categorical data. Then, we propose and describe a multi-objective optimization solution to obtain optimal categorical RBCs.
3.1. Rule-Based Classification for Categorical Data
Let be a classifier composed by M rules, where each rule , , has the following structure:
(2) |
where for the attribute (called antecedent) takes values in a set (), and (called consequent) takes values in (). Now, let be an observed example, with , for each . We propose maximum matching as reasoning method, where the compatibility degree of the rule for the example (denoted by ) is calculated as the number of attributes whose value coincides with that of the corresponding antecedent in , that is
where:
The association degree for the example with a class is computed by adding the compatibility degrees for the example of each rule whose consequent is equal to class c, that is:
where:
Therefore, the classification (or output) of the classifier for the example corresponds to the class whose association degree is maximum, that is:
3.2. A Multi-Objective Optimization Solution
Let be a dataset of K instances with p categorical input attributes, , and a categorical output attribute. Each input attribute j can take a category , , , and the output attribute can take a class , . The problem of finding an optimal classifier , as described in the previous section, can be formulated as an instance of the multi-objective constrained problem in Equation (1) with two objectives and two constraints:
(3) |
In the problem (Equation (3)), the function is a performance measure of the classifier over the dataset , the function is the number of rules of the classifier , and the constraints and limit the number of rules of the classifier to the interval , where w is the number of classes of the output attribute and is given by a user. Objectives and are in conflict. The fewer rules the classifier has, the fewer instances it can cover, that is, if the classifier is simpler it will have less capacity for prediction. There is, therefore, an intrinsic conflict between problem objectives (e.g., maximize accuracy and minimize model complexity) which cannot be easily aggregated to a single objective. Both objectives are typically optimized simultaneously in many other classification systems, such as neural networks or decision trees [47,48]. Figure 2 shows the Pareto front of a dummy binary classification problem described as in Equation (3), with rules, where is maximized. This front is composed of three non-dominated solutions (three possible classifiers) with two, three and four rules, respectively. The solutions with five and six rules are dominated (both by the solution with four rules).
Both ENORA and NSGA-II have been adapted to solve the problem described in Equation (3) with variable-length representation based on a Pittsburgh approach, uniform random initialization, binary tournament selection, handling constraints, ranking based on non-domination level with crowding distance, and self-adaptive variation operators. Self-adaptive variation operators work on different levels of the classifier: rule crossover, rule incremental crossover, rule incremental mutation, and integer mutation.
3.2.1. Representation
We use a variable-length representation based on a Pittsburgh approach [49], where each individual I of a population contains a variable number of rules , and each rule , is codified in the following components:
Integer values associated to the antecedents , for and .
Integer values associated to the consequent , for .
Additionally, to carry out self-adaptive crossing and mutation, each individual has two discrete parameters and associated with crossing and mutation, where is the number of crossing operators and is the number of mutation operators. Values and for self-adaptive variation are randomly generated from and , respectively. Table 1 summarizes the representation of an individual.
Table 1.
Codification for Rule Set | Codification for Adaptive Crossing and Mutation | |||||
---|---|---|---|---|---|---|
Antecedents | Consequent | Associated Crossing | Associated Mutation | |||
… | ||||||
⋮ | ⋮ | ⋮ | ⋮ | ⋮ | ||
… |
3.2.2. Constraint Handling
The constraints and are satisfied by means of specialized initialization and variation operators, which always generate individuals with a number of rules between w and .
3.2.3. Initial Population
The initial population (Algorithm 5) is randomly generated with the following conditions:
Individuals are uniformly distributed with respect to the number of rules with values between w and , and with an additional constraint that specifies that there must be at least one individual for each number of rules (Steps 4–8). This ensures an adequate initial diversity in the search space in terms of the second objective of the optimization model.
All individuals contain at least one rule for any output class between 1 and w (Steps 16–20).
Algorithm 5 Initialize population. |
Require: {Number of categorical input attributes} Require: , , {Number of categories for the input attributes} Require: , {Number of classes for the output attribute} Require: {Number of crossing operators} Require: {Number of mutation operators} Require: {Maximum number of rules} Require: {Number of individuals in the population}
|
3.2.4. Fitness Functions
Since the optimization model encompasses two objectives, each individual must be evaluated with two fitness functions, which correspond to the objective functions and of the problem (Equation (3)). The selection of the best individuals is done using the Pareto concept in a binary tournament.
3.2.5. Variation Operators
We use self-adaptive crossover and mutation, which means that the selection of the operators is made by means of an adaptive technique. As we have explained (cf. Section 3.2.1), each individual I has two integer parameters and to indicate which crossover or mutation is carried out. In our case, and are two crossover operators and two mutation operators, so that . Note that value 0 indicates that no crossover or no mutation is performed. Self-adaptive variation (Algorithm 6) generates two children from two parents by self-adaptive crossover (Algorithm 7) and self-adaptive mutation (Algorithm 8). Self-adaptive crossover of individuals and self-adaptive mutation of individual I are similar to each other. First, with a probability , the values and are replaced by a random value. Additionally, in the case of crossover, the value is replaced by . Then, the crossover indicated by or the mutation indicated by is performed. In summary, if an individual comes from a given crossover or a given mutation, that specific crossover and mutation are preserved to their offspring with probability , so the value of must be small enough to ensure a controlled evolution (in our case, we use ). Although the probability of the crossover and mutation is not explicitly represented, it can be computed as the ratio of the individuals for which crossover and mutation values are set to 1. As the population evolves, individuals with more successful types of crossover and mutation will be more common, so that the probability of selecting the more successful crossover and mutation types will increase. Using self-adaptive crossover and mutation operators helps to realize the goals of maintaining diversity in the population and sustaining the convergence capacity of the evolutionary algorithm, also eliminating the need of setting an a priori operator probability to each operator. In other approaches (e.g., [50]), the probabilities of crossover and mutation vary depending on the fitness value of the solutions.
Both ENORA and NSGA-II have been implemented with two crossover operators, rule crossover (Algorithm 9) and rule incremental crossover (Algorithm 10), and two mutation operators: rule incremental mutation (Algorithm 11) and integer mutation (Algorithm 12). Rule crossover randomly exchanges two rules selected from the parents, and rule incremental crossover adds to each parent a rule randomly selected from the other parent if its number of rules is less than the maximum number of rules. On the other hand, rule incremental mutation adds a new rule to the individual if the number of rules of the individual is less than the maximum number of rules, while integer mutation carries out a uniform mutation of a random antecedent belonging to a randomly selected rule.
Algorithm 6 Variation. |
Require:, {Individuals for variation}
|
Algorithm 7 Self-adaptive crossover. |
Require:I, J {Individuals for crossing} Require: () {Probability of variation} Require: {Number of different crossover operators}
|
Algorithm 8 Self-adaptive mutation. |
Require:I {Individual for mutation} Require: () {Probability of variation} Require: {Number of different mutation operators}
|
Algorithm 9 Rule crossover. |
Require:I, J {Individuals for crossing}
|
Algorithm 10 Rule incremental crossover. |
Require:I, J {Individuals for crossing} Require: {Maximum number of rules}
|
Algorithm 11 Rule incremental mutation. |
Require:I {Individual for mutation} Require: {Maximum number of rules}
|
Algorithm 12 Integer mutation. |
Require:I {Individual for mutation} Require: {Number of categorical input attributes} Require: , , {Number of categories for the input attributes}
|
4. Experiment and Results
To ensure the reproducibility of the experiments, we have used publicly available datasets. In particular, we have designed two sets of experiments, one using the Breast Cancer [51] dataset, and the other using the Monk’s Problem 2 [52] dataset.
4.1. The Breast Cancer Dataset
Breast Cancer encompasses 286 instances. Each instance corresponds to a patient who suffered from breast cancer and uses nine attributes to describe each patient. The class to be predicted is binary and represents whether the patient has suffered a recurring cancer event. In this dataset, 85 instances are positive and 201 are negative. Table 2 summarizes the attributes of the dataset. Among all instances, nine present some missing values; in the pre-processing phase, these have been replaced by the mode of the corresponding attribute.
Table 2.
# | Attribute Name | Type | Possible Values |
---|---|---|---|
1 | age | categorical | 10–19, 20–29, 30–39, 40–49, 50–59, 60–69, 70–79, 80–89, 90–99. |
2 | menopause | categorical | lt40, ge40, premeno |
3 | tumour-size | categorical | 0–4, 5–9, 10–14, 15–19, 20–24, 25–29, 30–34, 35–39, 40–44, 45–49, 50–54, 55–59 |
4 | inv-nodes | categorical | 0–2, 3–5, 6–8, 9–11, 12–14, 15–17, 18–20, 21–23, 24–26, 27–29, 30–32, 33–35, 36–39 |
5 | node-caps | categorical | yes, no |
6 | deg-malign | categorical | 1, 2, 3 |
7 | breast | categorical | left, right |
8 | breast-quad | categorical | left-up, left-low, right-up, right-low, central |
9 | irradiat | categorical | yes, no |
10 | class | categorical | no-recurrence-events, recurrence-events |
4.2. The Monk’s Problem 2 Dataset
In July 1991, the monks of Corsendonk Priory attended a summer course that was being held in their priory, namely the 2nd European Summer School on Machine Learning. After a week, the monks could not yet clearly identify the best ML algorithms, or which algorithms to avoid in which cases. For this reason, they decided to create the three so-called Monk’s problems, and used them to determine which ML algorithms were the best. These problems, rather simple and completely artificial, became later famous (because of their peculiar origin), and have been used as a comparison for many algorithms on several occasions. In particular, in [53], they have been used to test the performance of state-of-the-art (at that time) learning algorithms such as AQ17-DCI, AQ17-HCI, AQ17-FCLS, AQ14-NT, AQ15-GA, Assistant Professional, mFOIL, ID5R, IDL, ID5R-hat, TDIDT, ID3, AQR, CN2, WEB CLASS, ECOBWEB, PRISM, Backpropagation, and Cascade Correlation. For our research, we have used the Monk’s Problem 2, which contains six categorical input attributes and a binary output attribute, summarized in Table 3. The target concept associated with the Monk’s Problem 2 is the binary outcome of the logical formula:
Exactly two of: {heap_shape= round, body_shape=round, is_smiling=yes, holding=sword, jacket_color=red, has_tie=yes} |
In this dataset, the original training and testing sets were merged to allow other sampling procedures. The set contains a total of 601 instances, and no missing values.
Table 3.
# | Atttribute Name | Type | Possible Values |
---|---|---|---|
1 | head_shape | categorical | round, square, octagon |
2 | body_shape | categorical | round, square, octagon |
3 | is_smiling | categorical | yes, no |
4 | holding | categorical | sword, balloon, flag |
5 | jacket_color | categorical | red, yellow, green, blue |
6 | has_tie | categorical | yes, no |
7 | class | categorical | yes, no |
4.3. Optimization Models
We have conducted different experiments with different optimization models to calculate the overall performance of our proposed technique and to see the effect of optimizing different objectives for the same problem. First, we have designed a multi-objective constrained optimization model based on the accuracy:
(4) |
where is the proportion of correctly classified instances (both true positives and true negatives) among the total number of instances [54] obtained with the classifier for the dataset . is defined as:
where K is the number of instances of the dataset , and is the result of the classification of the instance i in with the classifier , that is:
where is the predicted value of the ith instance in , and is the corresponding true value in . Our second optimization model is based on the area under the ROC curve:
(5) |
where is the area under the ROC curve obtained with the classifier with the dataset . The ROC (Receiver Operating Characteristic) curve [55] is a graphical representation of the sensitivity versus the specificity index for a classifier varying the discrimination threshold value. Such a curve can be used to generate statistics that summarize the performance of a classifier, and it has been shown in [54] to be a simple, yet complete, empirical description of the decision threshold effect, indicating all possible combinations of the relative frequencies of the various kinds of correct and incorrect decisions. The area under the ROC curve can be computed as follows [56]:
where (sensitivity) is the proportion of positive instances classified as positive by the classifier in , (specificity) is the proportion of negative instances classified as negative by in , and t is the discrimination threshold. Finally, our third constrained optimization model is based on the root mean square error (RMSE):
(6) |
where is defined as the square root of the mean square error obtained with a classifier in the dataset :
where is the predicted value of the ith instance for the classifier , and is the corresponding output value in the database . Accuracy, area under the ROC curve, and root mean square error are all well-accepted measures used to evaluate the performance of a classifier. Therefore, it is natural to use such measures as fitting functions. In this way, we can establish which one behaves better in the optimization phase, and we can compare the results with those in the literature.
4.4. Choosing the Best Pareto Front
To compare the performance of ENORA and NSGA-II as metaheuristics in this particular optimization task, we use the hypervolume metric [57,58]. The hypervolume measures, simultaneously, the diversity and the optimality of the non-dominated solutions. The main advantage of using hypervolume against other standard measures, such as the error ratio, the generational distance, the maximum Pareto-optimal front error, the spread, the maximum spread, or the chi-square-like deviation, is that it can be computed without an optimal population, which is not always known [15]. The hypervolume is defined as the volume of the search space dominated by a population P, and is formulated as:
(7) |
where is the set of non-dominated individuals of P, and is the volume of the individual i. Subsequently, the hypervolume ratio (HVR) is defined as the ratio of the volume of the non-dominated search space over the volume of the entire search space, and is formulated as follows:
(8) |
where is the volume of the search space. Computing HVR requires reference points that identify the maximum and minimum values for each objective. For RBC optimization, as proposed in this work, the following minimum (, ) and maximum (, ) points, for each objective, are set in the multi-objective optimization models in Equations (4)–(6):
A first single execution of all six models (three driven by ENORA, and three driven by NSGA-II), over both datasets, has been designed for the purpose of showing the aspect of the final Pareto front, and compare the hypervolume ratio of the models. The results of this single execution, with population size equal to 50 and 20,000 generations (1,000,000 evaluations in total), are shown in Figure 3 and Figure 4 (by default, is set to 10, to which we add 2, because both datasets have a binary class). Regarding the configuration of the number of generations and the size of the population, our criterion has been established as follows: once the number of evaluations is set to 1,000,000, we can decide to use a population size of 100 individuals and 10,000 generations, or to use a population size of 50 individuals and 20,000 generations. The first configuration (100 × 10,000) allows a greater diversity with respect to the number of rules of the classifiers, while the second one (50 × 20,000) allows a better adjustment of the classifier parameters and therefore, a greater precision. Given the fact that the maximum number of rules of the classifiers is not greater than 12, we think that 50 individuals are sufficient to represent four classifiers on average for each number of rules (4 × 12 = 48∼50). Thus, we prefer the second configuration ( 20,000) because having more generations increases the chances of building classifiers with a higher precision.
Experiments were executed in a computer x64-based PC with one processor Intel64 Family 6 Model 60 Stepping 3 GenuineIntel 3201 Mhz, RAM 8131 MB. Table 4 shows the run time for each method over both datasets. Note that, although ENORA has less algorithmic complexity than NSGA-II, it has taken longer in experiments than NSGA-II. This is because the evaluation time of individuals in ENORA is higher than that of NSGA-II since ENORA has more diversity than NSGA-II, and therefore ENORA evaluates classifiers with more rules than NSGA-II.
Table 4.
Method | Breast Cancer | Monk’s Problem 2 |
---|---|---|
ENORA-ACC | 244.92 s. | 428.14 s. |
ENORA-AUC | 294.75 s. | 553.11 s. |
ENORA-RMSE | 243.30 s. | 414.42 s. |
NSGA-II-ACC | 127.13 s. | 260.83 s. |
NSGA-II-AUC | 197.07 s. | 424.83 s. |
NSGA-II-RMSE | 134.87 s. | 278.19 s. |
From these results, we can deduce that, first, ENORA maintains a higher diversity of the population, and achieves a better hypervolume ratio with respect to NSGA-II, and, second, using accuracy as the first objective generates better fronts than using the area under the ROC curve, which, in turn, performs better than using the root mean square error.
4.5. Comparing Our Method with Other Classifier Learning Systems (Full Training Mode)
To perform an initial comparison between the performance of the classifiers obtained with the proposed method and the ones obtained with classical methods (PART, JRip, OneR and ZeroR), we have executed again the six models in full training mode.
The parameters have been configured as in the previous experiment (population size equal to 50 and 20,000 generations), excepting the parameter that was set to 2 for the Breast Cancer dataset (this case), while, for the Monk’s Problem 2, it was set to 9. Observe that, since in both cases, executing the optimization models using leads to a single objective search for the Breast Cancer dataset. In fact, after the preliminary experiments were run, it turned out that the classical classifier learning systems tend to return very small, although not very precise, set of rules on Breast Cancer, and that justifies our choice. On the other hand, executing the classical rule learners on Monk’s Problem 2 returns more diverse sets of rules, which justifies choosing a higher in that case. To decide, a posteriori, which individual is chosen from the final front, we have used the default algorithm: the individual with the best value on the first objective is returned. In the case of Monk’s Problem 2, that individual has seven rules. The comparison is shown in Table 5 and Table 6, which show, for each classifier, the following information: number of rules, percent correct, true positive rate, false positive rate, precision, recall, F-measure, Matthews correlation coefficient, area under the ROC curve, area under precision-recall curve, and root mean square error. As for the Breast Cancer dataset (observe that the best result emerged from the proposed method), in the optimization model driven by NSGA-II, with root mean square error as the first objective (see Table 7), only PART was able to achieve similar results, although slightly worse, but at the price of having 15 rules, making the system clearly not interpretable. In the case of the Monk’s Problem 2 dataset, PART returned a model with 47 rules, which is not interpretable by any standard, although it is very accurate. The best interpretable result is the one with seven rules returned by ENORA, driven by the root mean square error (see Table 8). The experiments for classical learners have been conducted using the default parameters.
Table 5.
Learning Model | Number of Rules | Percent Correct | TP Rate | FP Rate | Precision | Recall | F-Measure | MCC | ROC Area | PRC Area | RMSE |
---|---|---|---|---|---|---|---|---|---|---|---|
ENORA-ACC | 2 | 79.02 | 0.790 | 0.449 | 0.796 | 0.790 | 0.762 | 0.455 | 0.671 | 0.697 | 0.458 |
ENORA-AUC | 2 | 75.87 | 0.759 | 0.374 | 0.751 | 0.759 | 0.754 | 0.402 | 0.693 | 0.696 | 0.491 |
ENORA-RMSE | 2 | 77.62 | 0.776 | 0.475 | 0.778 | 0.776 | 0.744 | 0.410 | 0.651 | 0.680 | 0.473 |
NSGA-II-ACC | 2 | 77.97 | 0.780 | 0.501 | 0.805 | 0.780 | 0.738 | 0.429 | 0.640 | 0.679 | 0.469 |
NSGA-II-AUC | 2 | 75.52 | 0.755 | 0.368 | 0.749 | 0.755 | 0.752 | 0.399 | 0.693 | 0.696 | 0.495 |
NSGA-II-RMSE | 2 | 79.37 | 0.794 | 0.447 | 0.803 | 0.794 | 0.765 | 0.467 | 0.673 | 0.700 | 0.454 |
PART | 15 | 78.32 | 0.783 | 0.397 | 0.773 | 0.783 | 0.769 | 0.442 | 0.777 | 0.793 | 0.398 |
JRip | 3 | 76.92 | 0.769 | 0.471 | 0.762 | 0.769 | 0.740 | 0.389 | 0.650 | 0.680 | 0.421 |
OneR | 1 | 72.72 | 0.727 | 0.563 | 0.703 | 0.727 | 0.680 | 0.241 | 0.582 | 0.629 | 0.522 |
ZeroR | - | 70.27 | 0.703 | 0.703 | 0.494 | 0.703 | 0.580 | 0.000 | 0.500 | 0.582 | 0.457 |
Table 6.
Learning Model | Number of Rules | Percent Correct | TP Rate | FP Rate | Precision | Recall | F-Measure | MCC | ROC Area | PRC Area | RMSE |
---|---|---|---|---|---|---|---|---|---|---|---|
ENORA-ACC | 7 | 75.87 | 0.759 | 0.370 | 0.753 | 0.759 | 0.745 | 0.436 | 0.695 | 0.680 | 0.491 |
ENORA-AUC | 7 | 68.71 | 0.687 | 0.163 | 0.836 | 0.687 | 0.687 | 0.523 | 0.762 | 0.729 | 0.559 |
ENORA-RMSE | 7 | 77.70 | 0.777 | 0.360 | 0.777 | 0.777 | 0.762 | 0.481 | 0.708 | 0.695 | 0.472 |
NSGA-II-ACC | 7 | 68.38 | 0.684 | 0.588 | 0.704 | 0.684 | 0.597 | 0.203 | 0.548 | 0.580 | 0.562 |
NSGA-II-AUC | 7 | 66.38 | 0.664 | 0.175 | 0.830 | 0.664 | 0.661 | 0.497 | 0.744 | 0.715 | 0.580 |
NSGA-II-RMSE | 7 | 68.71 | 0.687 | 0.591 | 0.737 | 0.687 | 0.595 | 0.226 | 0.548 | 0.583 | 0.559 |
PART | 47 | 94.01 | 0.940 | 0.087 | 0.940 | 0.940 | 0.940 | 0.866 | 0.980 | 0.979 | 0.218 |
JRip | 1 | 65.72 | 0.657 | 0.657 | 0.432 | 0.657 | 0.521 | 0.000 | 0.500 | 0.549 | 0.475 |
OneR | 1 | 65.72 | 0.657 | 0.657 | 0.432 | 0.657 | 0.521 | 0.000 | 0.500 | 0.549 | 0.585 |
ZeroR | - | 65.72 | 0.657 | 0.657 | 0.432 | 0.657 | 0.521 | 0.000 | 0.500 | 0.549 | 0.475 |
Table 7.
Rule | Antecedents | Consequent | ||||||
---|---|---|---|---|---|---|---|---|
: | IF | age = 50–59 | AND | inv-nodes = 0–2 | AND | node-caps = no | ||
AND | deg-malig = 1 | AND | breast = right | AND | breast-quad = left-low | THEN | class = no-recurrence-events | |
: | IF | age = 60–69 | AND | inv-nodes = 18–20 | AND | node-caps = yes | ||
AND | deg-malig = 3 | AND | breast = left | AND | breast-quad = right-up | THEN | class = recurrence-events |
Table 8.
Rule | Antecedents | Consequent | ||||||
---|---|---|---|---|---|---|---|---|
: | IF | head_shape = round | AND | body_shape = round | AND | is_smiling = no | ||
AND | holding = sword | AND | jacket_color = red | AND | has_tie = yes | THEN | class = yes | |
: | IF | head_shape = octagon | AND | body_shape = round | AND | is_smiling = no | ||
AND | holding = sword | AND | jacket_color = red | AND | has_tie = no | THEN | class = yes | |
: | IF | head_shape = round | AND | body_shape = round | AND | is_smiling = no | ||
AND | holding = sword | AND | jacket_color = yellow | AND | has_tie = yes | THEN | class = yes | |
: | IF | head_shape = round | AND | body_shape = round | AND | is_smiling = no | ||
AND | holding = sword | AND | jacket_color = red | AND | has_tie = no | THEN | class = yes | |
: | IF | head_shape = square | AND | body_shape = square | AND | is_smiling = yes | ||
AND | holding = flag | AND | jacket_color = yellow | AND | has_tie = no | THEN | class = no | |
: | IF | head_shape = octagon | AND | body_shape = round | AND | is_smiling = yes | ||
AND | holding = balloon | AND | jacket_color = blue | AND | has_tie = no | THEN | class = no | |
: | IF | head_shape = octagon | AND | body_shape = octagon | AND | is_smiling = yes | ||
AND | holding = sword | AND | jacket_color = green | AND | has_tie = no | THEN | class = no |
4.6. Comparing Our Method with Other Classifier Learning Systems (Cross-Validation and Train/Test Percentage Split Mode)
To test the capabilities of our methodology in a more significant way, we proceeded as follows. First, we designed a cross-validated experiment for the Breast Cancer dataset, in which we iterated three times a 10-fold cross-validation learning process [59] and considered the average value of the performance metrics percent correct, area under the ROC curve, and serialized model size of all results. Second, we designed a train/test percentage split experiment for the Monk’s Problem 2 dataset, in which we iterated ten times a 66% (training) versus 33% (testing) split and considered, again, the average result of the same metrics. Finally, we performed a statistical test over on results, to understand if they show any statistically significant difference. An execution of our methodology, and of standard classical learners, has been performed to obtain the models to be tested precisely under the same conditions of the experiment Section 4.5. It is worth observing that using two different types of evaluations allows us to make sure that our results are not influenced by the type of experiment. The results of the experiments are shown in Table 9 and Table 10.
Table 9.
Learning Model | Percent Correct | ROC Area | Serialized Model Size |
---|---|---|---|
ENORA-ACC | 73.45 | 0.61 | 9554.80 |
ENORA-AUC | 70.16 | 0.62 | 9554.63 |
ENORA-RMSE | 72.39 | 0.60 | 9557.77 |
NSGA-II-ACC | 72.50 | 0.60 | 9556.20 |
NSGA-II-AUC | 70.03 | 0.61 | 9555.70 |
NSGA-II-RMSE | 73.34 | 0.60 | 9558.60 |
PART | 68.92 | 0.61 | 55,298.13 |
JRip | 71.82 | 0.61 | 7664.07 |
OneR | 67.15 | 0.55 | 1524.00 |
ZeroR | 70.30 | 0.50 | 915.00 |
Table 10.
Learning Model | Percent Correct | ROC Area | Serialized Model Size |
---|---|---|---|
ENORA-ACC | 76.69 | 0.70 | 9586.50 |
ENORA-AUC | 72.82 | 0.79 | 9589.30 |
ENORA-RMSE | 75.66 | 0.68 | 9585.30 |
NSGA-II-ACC | 70.07 | 0.59 | 9590.60 |
NSGA-II-AUC | 67.08 | 0.70 | 9619.70 |
NSGA-II-RMSE | 67.63 | 0.54 | 9565.90 |
PART | 73.51 | 0.79 | 73,115.90 |
JRip | 64.05 | 0.50 | 5956.90 |
OneR | 65.72 | 0.50 | 1313.00 |
ZeroR | 65.72 | 0.50 | 888.00 |
The statistical tests aim to verify if there are significant differences among the means of each metric: percent correct, area under the ROC curve and serialized model size. We proceeded as follows. First, we checked normality and sphericity of each sample by means of the Shapiro–Wilk normality test. Then, if normality and sphericity conditions were met, we applied one way repeated measures ANOVA; otherwise, we applied the Friedman test. In the latter case, when statistically significant differences were detected, we applied the Nemenyi post-hoc test to locate where these differences were. Table A1, Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11 and Table A12 in Appendix A show the results of the performed tests for the Breast Cancer dataset for each of the three metrics, and Table A13, Table A14, Table A15, Table A16, Table A17, Table A18, Table A19, Table A20, Table A21, Table A22, Table A23 and Table A24 in Appendix B show the results for the Monk’s Problem 2 dataset.
4.7. Additional Experiments
Finally, we show the results of the evaluation with 10-fold cross-validation for Monk’s problem 2 dataset and for the following four other datasets:
Tic-Tac-Toe-Endgame dataset, with 9 input attributes, 958 instances, and binary class (Table 11).
Car dataset, with 6 input attributes, 1728 instances, and 4 output classes (Table 12).
Chess (King-Rook vs. King-Pawn) (kr-vs-kp), with 36 input attributes, 3196 instances, and binary class (Table 13).
Nursery dataset, with 8 input attributes, 12,960 instances, and 5 output classes (Table 14).
Table 11.
# | Attribute Name | Type | Possible Values |
---|---|---|---|
1 | top-left-square | categorical | x, o, b |
2 | top-middle-square | categorical | x, o, b |
3 | top-right-square | categorical | x, o, b |
4 | middle-left-square | categorical | x, o, b |
5 | middle-middle-square | categorical | x, o, b |
6 | middle-right-square | categorical | x, o, b |
7 | bottom-left-square | categorical | x, o, b |
8 | bottom-middle-square | categorical | x, o, b |
9 | bottom-right-square | categorical | x, o, b |
10 | class | categorical | positive, negative |
Table 12.
# | Attribute Name | Type | Possible Values |
---|---|---|---|
1 | buying | categorical | vhigh, high, med, low |
2 | maint | categorical | vhigh, high, med, low |
3 | doors | categorical | 2, 3, 4, 5-more |
4 | persons | categorical | 2, 4, more |
5 | lug_boot | categorical | small, med, big |
6 | safety | categorical | low, med, high |
7 | class | categorical | unacc, acc, good, vgood |
Table 13.
# | Attribute Name | Type | Possible Values |
---|---|---|---|
1 | bkblk | categorical | t, f |
2 | bknwy | categorical | t, f |
3 | bkon8 | categorical | t, f |
4 | bkona | categorical | t, f |
5 | bkspr | categorical | t, f |
6 | bkxbq | categorical | t, f |
7 | bkxcr | categorical | t, f |
8 | bkxwp | categorical | t, f |
9 | blxwp | categorical | t, f |
10 | bxqsq | categorical | t, f |
11 | cntxt | categorical | t, f |
12 | dsopp | categorical | t, f |
13 | dwipd | categorical | g, l |
14 | hdchk | categorical | t, f |
15 | katri | categorical | b, n, w |
16 | mulch | categorical | t, f |
17 | qxmsq | categorical | t, f |
18 | r2ar8 | categorical | t, f |
19 | reskd | categorical | t, f |
20 | reskr | categorical | t, f |
21 | rimmx | categorical | t, f |
22 | rkxwp | categorical | t, f |
23 | rxmsq | categorical | t, f |
24 | simpl | categorical | t, f |
25 | skach | categorical | t, f |
26 | skewr | categorical | t, f |
27 | skrxp | categorical | t, f |
28 | spcop | categorical | t, f |
29 | stlmt | categorical | t, f |
30 | thrsk | categorical | t, f |
31 | wkcti | categorical | t, f |
32 | wkna8 | categorical | t, f |
33 | wknck | categorical | t, f |
34 | wkovl | categorical | t, f |
35 | wkpos | categorical | t, f |
36 | wtoeg | categorical | n, t, f |
37 | class | categorical | won, nowin |
Table 14.
# | Attribute Name | Type | Possible Values |
---|---|---|---|
1 | parents | categorical | usual, pretentious, great_pret |
2 | has_nurs | categorical | proper, less_proper, improper, critical, very_crit |
3 | form | categorical | complete, completed, incomplete, foster |
4 | children | categorical | 1, 2, 3, more |
5 | housing | categorical | convenient, less_conv, critical |
6 | finance | categorical | convenient, inconv |
7 | social | categorical | nonprob, slightly_prob, problematic |
8 | health | categorical | recommended, priority, not_recom |
9 | class | categorical | not_recom, recommend, very_recom, priority, spec_prior |
We have used the ENORA algorithm together with the and objective functions in this case because these combinations have produced the best results for the Breast Cancer and Monk’s problem 2 datasets evaluated in 10-fold cross-validation (population size equal to 50, 20,000 generations and number of classes). Table 15 shows the results of the best combination ENORA-ACC or ENORA-RMSE together with the results of the classical rule-based classifiers.
Table 15.
Learning Model | Number of Rules | Percent Correct | TP Rate | FP Rate | Precision | Recall | F-Measure | MCC | ROC Area | PRC Area | RMSE |
---|---|---|---|---|---|---|---|---|---|---|---|
Monk’s problem 2 | |||||||||||
ENORA-ACC | 7 | 77.70 | 0.777 | 0.360 | 0.777 | 0.777 | 0.762 | 0.481 | 0.708 | 0.695 | 0.472 |
PART | 47 | 79.53 | 0.795 | 0.253 | 0.795 | 0.795 | 0.795 | 0.544 | 0.884 | 0.893 | 0.380 |
JRip | 1 | 62.90 | 0.629 | 0.646 | 0.526 | 0.629 | 0.535 | −0.034 | 0.478 | 0.537 | 0.482 |
OneR | 1 | 65.72 | 0.657 | 0.657 | 0.432 | 0.657 | 0.521 | 0.000 | 0.500 | 0.549 | 0.586 |
ZeroR | - | 65.72 | 0.657 | 0.657 | 0.432 | 0.657 | 0.521 | 0.000 | 0.491 | 0.545 | 0.457 |
Tic-Tac-Toe-Endgame | |||||||||||
ENORA-ACC/RMSE | 2 | 98.33 | 0.983 | 0.031 | 0.984 | 0.983 | 0.983 | 0.963 | 0.976 | 0.973 | 0.129 |
PART | 49 | 94.26 | 0.943 | 0.076 | 0.942 | 0.943 | 0.942 | 0.873 | 0.974 | 0.969 | 0.220 |
JRip | 9 | 97.81 | 0.978 | 0.031 | 0.978 | 0.978 | 0.978 | 0.951 | 0.977 | 0.977 | 0.138 |
OneR | 1 | 69.94 | 0.699 | 0.357 | 0.701 | 0.699 | 0.700 | 0.340 | 0.671 | 0.651 | 0.548 |
ZeroR | - | 65.35 | 0.653 | 0.653 | 0.427 | 0.653 | 0.516 | 0.000 | 0.496 | 0.545 | 0.476 |
Car | |||||||||||
ENORA-RMSE | 14 | 86.57 | 0.866 | 0.089 | 0.866 | 0.866 | 0.846 | 0.766 | 0.889 | 0.805 | 0.259 |
PART | 68 | 95.78 | 0.958 | 0.016 | 0.959 | 0.958 | 0.958 | 0.929 | 0.990 | 0.979 | 0.1276 |
JRip | 49 | 86.46 | 0.865 | 0.064 | 0.881 | 0.865 | 0.870 | 0.761 | 0.947 | 0.899 | 0.224 |
OneR | 1 | 70.02 | 0.700 | 0.700 | 0.490 | 0.700 | 0.577 | 0.000 | 0.500 | 0.543 | 0.387 |
ZeroR | - | 70.02 | 0.700 | 0.700 | 0.490 | 0.700 | 0.577 | 0.000 | 0.497 | 0.542 | 0.338 |
kr-vs-kp | |||||||||||
ENORA-RMSE | 10 | 94.87 | 0.949 | 0.050 | 0.950 | 0.949 | 0.949 | 0.898 | 0.950 | 0.927 | 0.227 |
PART | 23 | 99.06 | 0.991 | 0.010 | 0.991 | 0.991 | 0.991 | 0.981 | 0.997 | 0.996 | 0.088 |
JRip | 16 | 99.19 | 0.992 | 0.008 | 0.992 | 0.992 | 0.992 | 0.984 | 0.995 | 0.993 | 0.088 |
OneR | 1 | 66.46 | 0.665 | 0.350 | 0.675 | 0.665 | 0.655 | 0.334 | 0.657 | 0.607 | 0.579 |
ZeroR | - | 52.22 | 0.522 | 0.522 | 0.273 | 0.522 | 0.358 | 0.000 | 0.499 | 0.500 | 0.500 |
Nursery | |||||||||||
ENORA-ACC | 15 | 88.41 | 0.884 | 0.055 | 0.870 | 0.884 | 0.873 | 0.824 | 0.915 | 0.818 | 0.2153 |
PART | 220 | 99.21 | 0.992 | 0.003 | 0.992 | 0.992 | 0.992 | 0.989 | 0.999 | 0.997 | 0.053 |
JRip | 131 | 96.84 | 0.968 | 0.012 | 0.968 | 0.968 | 0.968 | 0.957 | 0.993 | 0.974 | 0.103 |
OneR | 1 | 70.97 | 0.710 | 0.137 | 0.695 | 0.710 | 0.702 | 0.570 | 0.786 | 0.632 | 0.341 |
ZeroR | - | 33.33 | 0.333 | 0.333 | 0.111 | 0.333 | 0.167 | 0.000 | 0.500 | 0.317 | 0.370 |
5. Analysis of Results and Discussion
The results of our tests allow for several considerations. The first interesting observation is that NSGA-II identifies fewer solutions than ENORA on the Pareto front, which implies less diversity and therefore a worse hypervolume ratio, as shown in Figure 3 and Figure 4. This is not surprising: in several other occasions [19,34,60], it has been shown that ENORA maintains a higher diversity in the population than other well-known evolutionary algorithms, with generally positive influence on the final results. Comparing the results in full training mode against the results in cross-validation or in splitting mode makes it evident that our solution produces classification models that are more resilient to over-fitting. For example, the classifier learned by PART with Monk’s Problem 2 presents a 94.01% accuracy in full training mode that drops to 73.51% in splitting mode. A similar, although with a more contained drop in accuracy, is shown by the classifier learned with Breast Cancer dataset; at the same time, the classifier learned by ENORA driven by accuracy shows only a 5.57% drop in one case, and even an improvement in the other case (see Table 5, Table 6, Table 9, and Table 10). This phenomenon is easily explained by looking at the number of rules: the more rules in a classifier, the higher the risk of over-fitting; PART produces very accurate classifiers, but at the price of adding many rules, which not only affects the interpretability of the model but also its resilience to over-fitting. Full training results seem to indicate that when the optimization model is driven by RMSE the classifiers are more accurate; nevertheless, they are also more prone to over-fitting, indicating that, on average, the optimization models driven by the accuracy are preferable.
From the statistical tests (whose results are shown in the Appendix A and Appendix B) we conclude that among the six variants of the proposed optimization model there are no statistical significative differences, which suggests that the advantages of our method do not depend directly on a specific evolutionary algorithm or on the specific performance measure that is used to drive the evolutions. Significant statistical differences between our method and very simple classical methods such as OneR were expectable. Significant statistical differences between our method and a well-consolidated one such as PART have not been found, but the price to be paid for using PART in order to have similar results to ours is a very high number of rules (15 vs. 2 in one case and 47 vs. 7 in the other case).
We would like to highlight that both the Breast Cancer dataset and the Monk’s problem 2 dataset are difficult to approximate with interpretable classifiers and that none of the analyzed classifiers obtains high accuracy rates using the cross-validation technique. Even powerful black-box classifiers, such as Random Forest and Logistic, obtain success rates below in 10-fold cross-validation for these datasets. However, ENORA obtains a better balance (trade-off) between precision and interpretability than the rest of the classifiers. For the rest of the analyzed datasets, the accuracy obtained using ENORA is substantially higher. For example, for the Tic-Tac-Toe-Endgame dataset, ENORA obtains a success percentage with only two rules in cross-validation, while PART obtains with 49 rules, and JRip obtains with nine rules. With respect to the results obtained in the datasets Car, kr-vs-kp and Nursery, we want to comment that better success percentage can be obtained if the maximum number of evaluations is increased. However, better success percentages imply a greater number of rules, which is to the detriment of the interpretability of the models.
6. Conclusions and Future Works
In this paper, we have proposed a novel technique for categorical classifier learning. Our proposal is based on defining the problem of learning a classifier as a multi-objective optimization problem, and solving it by suitably adapting an evolutionary algorithm to this task; our two objectives are minimizing the number of rules (for a better interpretability of the classifier) and maximizing a metric of performance. Depending on the particular metric that is chosen, (slightly) different optimization models arise. We have tested our proposal, in a first instance, on two different publicly available datasets, Breast Cancer (in which each instance represents a patient that has suffered from breast cancer and is described by nine attributes, and the class to be predicted represents the fact that the patient has suffered a recurring event) and Monk’s Problem 2 (which is an artificial, well-known dataset in which the class to be predicted represents a logical function), using two different evolutionary algorithms, namely ENORA and NSGA-II, and three different choices as a performance metric, i.e., accuracy, the area under the ROC curve, and the root mean square error. Additionally, we have shown the results of the evaluation in 10-fold cross-validation of the publicly available Tic-Tac-Toe-Endgame, Car, kr-vs-kp and Nursery datasets.
Our initial motivation was to design a classifier learning system that produces interpretable, yet accurate, classifiers: since interpretability is a direct function of the number of rules, we conclude that such an objective has been achieved. As an aside, observe that our approach allows the user to decide, beforehand, a maximum number of rules; this can also be done in PART and JRip, but only indirectly. Finally, the idea underlying our approach is that multiple classifiers are explored at the same time in the same execution, and this allows us to choose the best compromise between the performance and the interpretability of a classifier a posteriori.
As a future work, we envisage that our methodology can benefit from an embedded future selection mechanism. In fact, all attributes are (ideally) used in every rule of a classifier learned by our optimization model. By simply relaxing such a constraint, and by suitably re-defining the first objective in the optimization model (e.g., by minimizing the sum of the lengths of all rules, or similar measures), the resulting classifiers will naturally present rules that use more features as well as rules that use less (clearly, the implementation must be adapted to obtain an initial population in which the classifiers have rules of different lengths as well as mutation operators that allow a rule to grow or to shrink). Although this approach does not follow the classical definition of feature selection mechanisms (in which a subset of features is selected that reduces the dataset over which a classifier is learned), it is natural to imagine that it may produce even more accurate classifiers, and more interpretable at the same time.
Currently, we are implementing our own version of multi-objective differential evolution (MODE) for rule-based classification for inclusion in the Weka Open Source Software issued under the GNU General Public License. The implementation of other algorithms, such as MOEA/D, their adaptation in the Weka development platform and subsequent analysis and comparison are planned for future work.
Acknowledgments
This study was partially supported by computing facilities of Extremadura Research Centre for Advanced Technologies (CETA-CIEMAT), funded by the European Regional Development Fund (ERDF). CETA-CIEMAT belongs to CIEMAT and the Government of Spain.
Abbreviations
The following abbreviations are used in this manuscript:
ML | Machine learning |
ANN | Artificial neural networks |
DLNN | Deep learning neural networks |
CEO | Chief executive officer |
SVM | Support vector machines |
IBL | Instance-based learning |
DT | Decision trees |
RBC | Rule-based classifiers |
ROC | Receiver operating characteristic |
RMSE | Root mean square error performance metric |
FL | Fuzzy logic |
MOEA | Multi-objective evolutionary algorithms |
NSGA-II | Non-dominated sorting genetic algorithm, 2nd version |
ENORA | Evolutionary non-dominated radial slots based algorithm |
PART | Partial decision tree classifier |
JRip | RIPPER classifier of Weka |
RIPPER | Repeated incremental pruning to produce error reduction |
OneR | One rule classifier |
ZeroR | Zero rule classifier |
ENORA-ACC | ENORA with objective function defined as accuracy |
ENORA-AUC | ENORA with objective function defined as area under the ROC curve |
ENORA-RMSE | ENORA with RMSE objective function |
NSGA-II-ACC | NSGA-II with objective function defined as accuracy |
NSGA-II-AUC | NSGA-II with objective function defined as area under the ROC curve |
NSGA-II-RMSE | NSGA-II with RMSE objective function |
HVR | Hypervolume ratio |
TP | True positive |
FP | False positive |
MCC | Matthews correlation coefficient |
PRC | Precision-recall curve |
Appendix A. Statistical Tests for Breast Cancer Dataset
Table A1.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 0.5316 | Not Rejected |
ENORA-AUC | 0.3035 | Not Rejected |
ENORA-RMSE | 0.7609 | Not Rejected |
NSGA-II-ACC | 0.1734 | Not Rejected |
NSGA-II-AUC | 0.3802 | Not Rejected |
NSGA-II-RMSE | 0.6013 | Not Rejected |
PART | 0.0711 | Not Rejected |
JRip | 0.5477 | Not Rejected |
OneR | 0.316 | Not Rejected |
ZeroR | 3.818 × 10 | Rejected |
Table A2.
p-Value | Null Hypothesis | |
---|---|---|
Friedman | 5.111 × 10 | Rejected |
Table A3.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 0.2597 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 0.9627 | 0.9627 | - | - | - | - | - | - | - |
NSGA-II-ACC | 0.9981 | 0.8047 | 1.0000 | - | - | - | - | - | - |
NSGA-II-AUC | 0.2951 | 1.0000 | 0.9735 | 0.8386 | - | - | - | - | - |
NSGA-II-RMSE | 1.0000 | 0.2169 | 0.9436 | 0.9960 | 0.2486 | - | - | - | - |
PART | 0.1790 | 1.0000 | 0.9186 | 0.6997 | 1.0000 | 0.1461 | - | - | - |
JRip | 0.9909 | 0.8956 | 1.0000 | 1.0000 | 0.9186 | 0.9840 | 0.8164 | - | - |
OneR | 0.0004 | 0.6414 | 0.0451 | 0.0108 | 0.5961 | 0.0002 | 0.7546 | 0.0212 | - |
ZeroR | 0.2377 | 1.0000 | 0.9538 | 0.7803 | 1.0000 | 0.1973 | 1.0000 | 0.8783 | 0.6709 |
Table A4.
ENORA-ACC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-RMSE | JRip | |
---|---|---|---|---|---|
OneR | ENORA-ACC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-RMSE | JRip |
Table A5.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 0.6807 | Not Rejected |
ENORA-AUC | 0.3171 | Not Rejected |
ENORA-RMSE | 0.6125 | Not Rejected |
NSGA-II-ACC | 0.0871 | Not Rejected |
NSGA-II-AUC | 0.5478 | Not Rejected |
NSGA-II-RMSE | 0.6008 | Not Rejected |
PART | 0.6066 | Not Rejected |
JRip | 0.2978 | Not Rejected |
OneR | 0.4531 | Not Rejected |
ZeroR | 0.0000 | Rejected |
Table A6.
p-Value | Null Hypothesis | |
---|---|---|
Friedman | 8.232 × 10 | Rejected |
Table A7.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 1.0000 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 0.9972 | 0.9990 | - | - | - | - | - | - | - |
NSGA-II-ACC | 0.9999 | 1.0000 | 1.0000 | - | - | - | - | - | - |
NSGA-II-AUC | 1.0000 | 1.0000 | 1.0000 | 1.0000 | - | - | - | - | - |
NSGA-II-RMSE | 0.9990 | 0.9997 | 1.0000 | 1.0000 | 1.0000 | - | - | - | - |
PART | 0.9999 | 1.0000 | 1.0000 | 1.0000 | 1.0000 | 1.0000 | - | - | - |
JRip | 1.0000 | 1.0000 | 0.9992 | 1.0000 | 1.0000 | 0.9998 | 1.0000 | - | - |
OneR | 0.0041 | 0.0062 | 0.0790 | 0.0323 | 0.0281 | 0.0582 | 0.0345 | 0.0067 | - |
ZeroR | 3.8 × 10 | 7.2 × 10 | 4.6 × 10 | 9.8 × 10 | 7.8 × 10 | 2.7 × 10 | 1.1 × 10 | 8.1 × 10 | 0.6854 |
Table A8.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | |
---|---|---|---|---|---|---|---|---|
OneR | ENORA-ACC | ENORA-AUC | - | NSGA-II-ACC | NSGA-II-AUC | - | PART | JRip |
ZeroR | ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip |
Table A9.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 5.042 × 10 | Rejected |
ENORA-AUC | 2.997 × 10 | Rejected |
ENORA-RMSE | 4.762 × 10 | Rejected |
NSGA-II-ACC | 4.88 × 10 | Rejected |
NSGA-II-AUC | 2.339 × 10 | Rejected |
NSGA-II-RMSE | 2.708 × 10 | Rejected |
PART | 0.3585 | Not Rejected |
JRip | 9.086 × 10 | Rejected |
OneR | 1.007 × 10 | Rejected |
ZeroR | 0.0000 | Rejected |
Table A10.
p-Value | Null Hypothesis | |
---|---|---|
Friedman | 2.2 × 10 | Rejected |
Table A11.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 0.9998 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 0.0053 | 0.0004 | - | - | - | - | - | - | - |
NSGA-II-ACC | 0.3871 | 0.0942 | 0.8872 | - | - | - | - | - | - |
NSGA-II-AUC | 0.8872 | 0.4894 | 0.3871 | 0.9988 | - | - | - | - | - |
NSGA-II-RMSE | 4.1 × 10 | 1.3 × 10 | 0.9860 | 0.2169 | 0.0244 | - | - | - | - |
PART | 4.7 × 10 | 5.6 × 10 | 0.1973 | 0.0013 | 3.3 × 10 | 0.8689 | - | - | - |
JRip | 0.2712 | 0.6997 | 1.2 × 10 | 7.0 × 10 | 0.0025 | 6.3 × 10 | 6.9 × 10 | - | - |
OneR | 0.0062 | 0.0546 | 1.5 × 10 | 5.5 × 10 | 5.5 × 10 | 8.3 × 10 | 8.3 × 10 | 0.9584 | - |
ZeroR | 1.9 × 10 | 0.0004 | 7.3 × 10 | 8.6 × 10 | 2.3 × 10 | 8.5 × 10 | <2 × 10 | 0.2377 | 0.9584 |
Table A12.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | |
---|---|---|---|---|---|---|---|
ENORA-RMSE | ENORA-ACC | NSGA-II-AUC | - | - | - | - | - |
NSGA-II-RMSE | ENORA-ACC | ENORA-AUC | - | - | NSGA-II-AUC | - | - |
PART | ENORA-ACC | ENORA-AUC | - | NSGA-II-ACC | NSGA-II-AUC | - | - |
JRip | - | - | JRip | JRip | JRip | JRip | JRip |
OneR | OneR | - | OneR | OneR | OneR | OneR | OneR |
ZeroR | ZeroR | ZeroR | ZeroR | ZeroR | ZeroR | ZeroR | ZeroR |
Appendix B. Statistical Tests for Monk’s Problem 2 Dataset
Table A13.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 0.6543 | Not Rejected |
ENORA-AUC | 0.6842 | Not Rejected |
ENORA-RMSE | 0.0135 | Rejected |
NSGA-II-ACC | 0.979 | Not Rejected |
NSGA-II-AUC | 0.382 | Not Rejected |
NSGA-II-RMSE | 0.0486 | Rejected |
PART | 0.5671 | Not Rejected |
JRip | 0.075 | Rejected |
OneR | 4.672 × 10 | Rejected |
ZeroR | 4.672 × 10 | Rejected |
Table A14.
p-Value | Null Hypothesis | |
---|---|---|
Frideman | 1.292 × 10 | Rejected |
Table A15.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 0.8363 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 1.0000 | 0.9471 | - | - | - | - | - | - | - |
NSGA-II-ACC | 0.1907 | 0.9902 | 0.3481 | - | - | - | - | - | - |
NSGA-II-AUC | 0.0126 | 0.6294 | 0.0342 | 0.9958 | - | - | - | - | - |
NSGA-II-RMSE | 0.0126 | 0.6294 | 0.0342 | 0.9958 | 1.0000 | - | - | - | - |
PART | 0.8714 | 1.0000 | 0.9631 | 0.9841 | 0.5769 | 0.5769 | - | - | - |
JRip | 2.1 × 10 | 0.0048 | 1.0 × 10 | 0.1341 | 0.6806 | 0.6806 | 0.0036 | - | - |
OneR | 0.0001 | 0.0743 | 0.0006 | 0.6032 | 0.9875 | 0.9875 | 0.0601 | 0.9984 | - |
ZeroR | 0.0001 | 0.0743 | 0.0006 | 0.6032 | 0.9875 | 0.9875 | 0.0601 | 0.9984 | 1.0000 |
Table A16.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | PART | |
---|---|---|---|---|
NSGA-II-AUC | ENORA-ACC | - | ENORA-RMSE | - |
NSGA-II-RMSE | ENORA-ACC | - | ENORA-RMSE | - |
JRip | ENORA-ACC | ENORA-AUC | ENORA-RMSE | PART |
OneR | ENORA-ACC | - | ENORA-RMSE | - |
ZeroR | ENORA-ACC | - | ENORA-RMSE | - |
Table A17.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 0.4318 | Not Rejected |
ENORA-AUC | 0.7044 | Not Rejected |
ENORA-RMSE | 0.0033 | Rejected |
NSGA-II-ACC | 0.3082 | Not Rejected |
NSGA-II-AUC | 0.0243 | Rejected |
NSGA-II-RMSE | 0.7802 | Not Rejected |
PART | 0.1641 | Not Rejected |
JRip | 0.3581 | Not Rejected |
OneR | 0.0000 | Rejected |
ZeroR | 0.0000 | Rejected |
Table A18.
p-Value | Null Hypothesis | |
---|---|---|
Frideman | 1.051 × 10 | Rejected |
Table A19.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 0.8363 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 1.0000 | 0.7054 | - | - | - | - | - | - | - |
NSGA-II-ACC | 0.8870 | 0.0539 | 0.9556 | - | - | - | - | - | - |
NSGA-II-AUC | 1.0000 | 0.8544 | 1.0000 | 0.8713 | - | - | - | - | - |
NSGA-II-RMSE | 0.5504 | 0.0084 | 0.7054 | 0.9999 | 0.5239 | - | - | - | - |
PART | 0.7054 | 1.0000 | 0.5504 | 0.0269 | 0.7295 | 0.0036 | - | - | - |
JRip | 0.0238 | 2.3 × 10 | 0.0482 | 0.6806 | 0.0211 | 0.9471 | 7.0 × 10 | - | - |
OneR | 0.0084 | 4.7 × 10 | 0.0186 | 0.4715 | 0.0073 | 0.8363 | 1.4 × 10 | 1.0000 | - |
ZeroR | 0.0084 | 4.7 × 10 | 0.0186 | 0.4715 | 0.0073 | 0.8363 | 1.4 × 10 | 1.0000 | 1.0000 |
Table A20.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | PART | ||
---|---|---|---|---|---|---|---|
NSGA-II-RMSE | - | ENORA-AUC | - | - | - | - | - |
PART | - | - | - | PART | - | PART | - |
JRip | ENORA-ACC | ENORA-AUC | ENORA-RMSE | - | NSGA-II-AUC | - | PART |
OneR | ENORA-ACC | ENORA-AUC | ENORA-RMSE | - | NSGA-II-AUC | - | PART |
ZeroR | ENORA-ACC | ENORA-AUC | ENORA-RMSE | - | NSGA-II-AUC | - | PART |
Table A21.
Algorithm | p-Value | Null Hypothesis |
---|---|---|
ENORA-ACC | 4.08 × 10 | Rejected |
ENORA-AUC | 0.0002 | Rejected |
ENORA-RMSE | 0.0094 | Rejected |
NSGA-II-ACC | 0.0192 | Rejected |
NSGA-II-AUC | 0.0846 | Rejected |
NSGA-II-RMSE | 0.0037 | Rejected |
PART | 0.9721 | Not Rejected |
JRip | 0.0068 | Rejected |
OneR | 0.0000 | Rejected |
ZeroR | 0.0000 | Rejected |
Table A22.
p-Value | Null Hypothesis | |
---|---|---|
Frideman | 2.657 × 10 | Rejected |
Table A23.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | JRip | OneR | |
---|---|---|---|---|---|---|---|---|---|
ENORA-AUC | 1.0000 | - | - | - | - | - | - | - | - |
ENORA-RMSE | 1.0000 | 1.0000 | - | - | - | - | - | - | - |
NSGA-II-ACC | 1.0000 | 1.0000 | 1.0000 | - | - | - | - | - | - |
NSGA-II-AUC | 0.9925 | 0.9696 | 0.9984 | 0.9841 | - | - | - | - | - |
NSGA-II-RMSE | 0.8870 | 0.9556 | 0.7966 | 0.9267 | 0.2622 | - | - | - | - |
PART | 0.2824 | 0.1752 | 0.3957 | 0.2246 | 0.9015 | 0.0027 | - | - | - |
JRip | 0.1752 | 0.2824 | 0.1110 | 0.2246 | 0.0084 | 0.9752 | 1.0 × 10 | - | - |
OneR | 0.0211 | 0.0431 | 0.0110 | 0.0304 | 0.0004 | 0.6552 | 1.5 × 10 | 0.9993 | - |
ZeroR | 0.0012 | 0.0031 | 0.0006 | 0.0020 | 1.0 × 10 | 0.1907 | 1.3 × 10 | 0.9015 | 0.9993 |
Table A24.
ENORA-ACC | ENORA-AUC | ENORA-RMSE | NSGA-II-ACC | NSGA-II-AUC | NSGA-II-RMSE | PART | |
---|---|---|---|---|---|---|---|
PART | - | - | - | - | - | NSGA-II-RMSE | - |
JRip | - | - | - | - | JRip | - | JRip |
OneR | OneR | OneR | OneR | OneR | OneR | - | OneR |
ZeroR | ZeroR | ZeroR | ZeroR | ZeroR | ZeroR | - | ZeroR |
Appendix C. Nomenclature
Table A25.
Symbol | Definition |
---|---|
Equation (1): Multi-objective constrained optimization | |
k-th decision variable | |
Set of decision variables | |
i-th objective function | |
j-th constraint | |
Number of objectives | |
Number of constraints | |
Number of decision variables | |
Domain for each each decision variable | |
Domain for the set of decision variables | |
Set of all feasible solutions | |
Set of non-dominated solutions or Pareto optimal set | |
Pareto domination function | |
Equation (2): Rule-based classification for categorical data | |
Dataset | |
categorical input attribute in the dataset | |
Categorical input attributes in the dataset | |
y | Categorical output attribute in the dataset |
Domain of i-th categorical input attribute in the dataset | |
Domain of categorical output attribute in the dataset | |
Number of categorical input attributes in the dataset | |
Rule-based classifier | |
rule of classifier | |
Category for categorical input attribute and rule of classifier | |
Category for categorical output attribute and rule of classifier | |
Compatibility degree of the rule of classifier for the example | |
Result of the rule of classifier and categorical input attribute | |
Association degree of classifier for the example with the class c | |
Result of of the rule of classifier for the example with the class c | |
Classification or output of the classifier for the example | |
Equation (3): Multi-objective constrained optimization problem for rule-based classification | |
Performance objective function of the classifier in the dataset | |
Number of rules of the classifier | |
Maximum number of rules allowed for classifiers | |
Equations (4)–(6): Optimization models | |
Acurracy: proportion of correctly classified instances with the classifier in the dataset | |
K | Number of instances in the dataset |
Result of the classification of the instance in the dataset with the classifier | |
Predicted value of the instance in the dataset with the classifier | |
Corresponding true value for the instance in the dataset . | |
Area under the ROC curve obtained with the classifier in the dataset . | |
Sensitivity: proportion of positive instances classified as positive with the classifier in the dataset | |
Specificity: proportion of negative instances classified as negative with the classifier in the dataset | |
t | Discrimination threshold |
Square root of the mean square error obtained with the classifier in the dataset |
Table A26.
Equations (7) and (8): Hypervolume metric | |
P | Population |
Set of non-dominated individuals of P | |
Volume of the search space dominated by the individual i | |
Hypervolume: volume of the search space dominated by population P | |
Volume of the search space non-dominated by population P | |
Hypervolume ratio: ratio of over the volume of the entire search space | |
Volume of the search space | |
Minimum value for objective | |
Maximum value for objective | |
Minimum value for objective | |
Maximum value for objective |
Author Contributions
Conceptualization, F.J. and G.S. (Gracia Sánchez); Methodology, F.J. and G.S. (Guido Sciavicco); Software, G.S. (Gracia Sánchez) and C.M.; Validation, F.J., G.S. (Gracia Sánchez) and C.M.; Formal Analysis, F.J. and G.S. (Guido Sciavicco); Investigation, F.J. and G.S. (Gracia Sánchez); Resources, L.M.; Data Curation, L.M.; Writing—Original Draft Preparation, F.J., L.M. and G.S. (Guido Sciavicco); Writing—Review and Editing, F.J., L.M. and G.S. (Guido Sciavicco); Visualization, F.J.; Supervision, F.J.; Project Administration, F.J.; and Funding Acquisition, F.J., L.M., G.S. (Gracia Sánchez) and G.S. (Guido Sciavicco).
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Bishop C.M. Pattern Recognition and Machine Learning (Information Science and Statistics) Springer; Berlin/Heidelberg, Germany: 2006. [Google Scholar]
- 2.Russell S., Norvig P. Artificial Intelligence: A Modern Approach. 3rd ed. Prentice Hall Press; Upper Saddle River, NJ, USA: 2009. [Google Scholar]
- 3.Davalo É. Neural Networks. Macmillan Education; London, UK: 1991. MacMillan Computer Science. [Google Scholar]
- 4.Goodfellow I., Bengio Y., Courville A. Deep Learning. MIT Press; Cambridge, MA, USA: 2016. [Google Scholar]
- 5.Cortes C., Vapnik V. Support-vector networks. Mach. Learn. 1995;20:273–297. doi: 10.1007/BF00994018. [DOI] [Google Scholar]
- 6.Aha D.W., Kibler D., Albert M.K. Instance-based learning algorithms. Mach. Learn. 1991;6:37–66. doi: 10.1007/BF00153759. [DOI] [Google Scholar]
- 7.Gacto M., Alcalá R., Herrera F. Interpretability of linguistic fuzzy rule-based systems: An overview of interpretability measures. Inf. Sci. 2011;181:4340–4360. doi: 10.1016/j.ins.2011.02.021. [DOI] [Google Scholar]
- 8.Cano A., Zafra A., Ventura S. An EP algorithm for learning highly interpretable classifiers; Proceedings of the 11th International Conference on Intelligent Systems Design and Applications; Cordoba, Spain. 22–24 November 2011; pp. 325–330. [Google Scholar]
- 9.Liu H., Gegov A. Collaborative Decision Making by Ensemble Rule Based Classification Systems. In: Pedrycz W., Chen S.-M., editors. Granular Computing and Decision-Making: Interactive and Iterative Approaches. Springer International Publishing; Cham, Switzerland: 2015. pp. 245–264. [Google Scholar]
- 10.Sulzmann J.N., Fürnkranz J. In: Rule Stacking: An Approach for Compressing an Ensemble of Rule Sets into a Single Classifier. Elomaa T., Hollmén J., Mannila H., editors. Springer; Berlin/Heidelberg, Germany: 2011. pp. 323–334. Discovery Science. [Google Scholar]
- 11.Jin Y. Fuzzy Modeling of High-Dimensional Systems: Complexity Reduction and Interpretability Improvement. IEEE Trans. Fuzzy Syst. 2000;8:212–220. doi: 10.1109/91.842154. [DOI] [Google Scholar]
- 12.Breiman L., Friedman J.H., Olshen R.A., Stone C.J. Classification and Regression Trees. Wadsworth and Brooks; Monterey, CA, USA: 1984. [Google Scholar]
- 13.Novák V., Perfilieva I., Mockor J. Mathematical Principles of Fuzzy Logic. Springer Science + Business Media; Heidelberg, Germany: 2012. [Google Scholar]
- 14.Freund Y., Schapire R.E. A Short Introduction to Boosting; Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence; Stockholm, Sweden. 31 July–6 August 1999; pp. 1401–1406. [Google Scholar]
- 15.Deb K. Multi-Objective Optimization Using Evolutionary Algorithms. John Wiley and Sons; London, UK: 2001. [Google Scholar]
- 16.Coello C.A.C., van Veldhuizen D.A., Lamont G.B. Evolutionary Algorithms for Solving Multi-Objective Problems. Kluwer Academic/Plenum Publishers; New York, NY, USA: 2002. [Google Scholar]
- 17.Jiménez F., Gómez-Skarmeta A., Sánchez G., Deb K. An evolutionary algorithm for constrained multi-objective optimization; Proceedings of the 2002 Congress on Evolutionary Computation; Honolulu, HI, USA. 12–17 May 2002; pp. 1133–1138. [Google Scholar]
- 18.Jiménez F., Sánchez G., Juárez J.M. Multi-objective evolutionary algorithms for fuzzy classification in survival prediction. Artif. Intell. Med. 2014;60:197–219. doi: 10.1016/j.artmed.2013.12.006. [DOI] [PubMed] [Google Scholar]
- 19.Jiménez F., Marzano E., Sánchez G., Sciavicco G., Vitacolonna N. Attribute selection via multi-objective evolutionary computation applied to multi-skill contact center data classification; Proceedings of the 2015 IEEE Symposium Series on Computational Intelligence; Cape Town, South Africa. 7–10 December 2015; pp. 488–495. [Google Scholar]
- 20.Jiménez F., Jódar R., del Pilar Martín M., Sánchez G., Sciavicco G. Unsupervised feature selection for interpretable classification in behavioral assessment of children. Expert Syst. 2017;34:e12173. doi: 10.1111/exsy.12173. [DOI] [Google Scholar]
- 21.Rey M., Galende M., Fuente M., Sainz-Palmero G. Multi-objective based Fuzzy Rule Based Systems (FRBSs) for trade-off improvement in accuracy and interpretability: A rule relevance point of view. Knowl.-Based Syst. 2017;127:67–84. doi: 10.1016/j.knosys.2016.12.028. [DOI] [Google Scholar]
- 22.Ducange P., Lazzerini B., Marcelloni F. Multi-objective genetic fuzzy classifiers for imbalanced and cost-sensitive datasets. Soft Comput. 2010;14:713–728. doi: 10.1007/s00500-009-0460-y. [DOI] [Google Scholar]
- 23.Gorzalczany M.B., Rudzinski F. A multi-objective genetic optimization for fast, fuzzy rule-based credit classification with balanced accuracy and interpretability. Appl. Soft Comput. 2016;40:206–220. doi: 10.1016/j.asoc.2015.11.037. [DOI] [Google Scholar]
- 24.Ducange P., Mannara G., Marcelloni F., Pecori R., Vecchio M. A novel approach for internet traffic classification based on multi-objective evolutionary fuzzy classifiers; Proceedings of the 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE); Naples, Italy. 9–12 July 2017; pp. 1–6. [Google Scholar]
- 25.Antonelli M., Bernardo D., Hagras H., Marcelloni F. Multiobjective Evolutionary Optimization of Type-2 Fuzzy Rule-Based Systems for Financial Data Classification. IEEE Trans. Fuzzy Syst. 2017;25:249–264. doi: 10.1109/TFUZZ.2016.2578341. [DOI] [Google Scholar]
- 26.Carmona C.J., González P., Deljesus M.J., Herrera F. NMEEF-SD: Non-dominated multiobjective evolutionary algorithm for extracting fuzzy rules in subgroup discovery. IEEE Trans. Fuzzy Syst. 2010;18:958–970. doi: 10.1109/TFUZZ.2010.2060200. [DOI] [Google Scholar]
- 27.Hubertus T., Klaus M., Eberhard T. Optimization Theory. Kluwer Academic Publishers; Dordrecht, The Netherlands: 2004. [Google Scholar]
- 28.Sinha S. Mathematical Programming: Theory and Methods. Elsevier; Amsterdam, The Netherlands: 2006. [Google Scholar]
- 29.Collette Y., Siarry P. Multiobjective Optimization: Principles and Case Studies. Springer-Verlag Berlin Heidelberg; New York, NY, USA: 2004. [Google Scholar]
- 30.Karloff H. Linear Programming. Birkhauser Basel; Boston, MA, USA: 1991. [Google Scholar]
- 31.Maros I., Mitra G. Simplex algorithms. In: Beasley J.E., editor. Advances in Linear and Integer Programming. Oxford University Press; Oxford, UK: 1996. pp. 1–46. [Google Scholar]
- 32.Bertsekas D. Nonlinear Programming. 2nd ed. Athena Scientific; Cambridge, MA, USA: 1999. [Google Scholar]
- 33.Jiménez F., Verdegay J.L. Computational Intelligence in Theory and Practice. In: Reusch B., Temme K.-H., editors. Advances in Soft Computing. Springer; Heidelberg, Germany: 2001. pp. 167–182. [Google Scholar]
- 34.Jiménez F., Sánchez G., García J., Sciavicco G., Miralles L. Multi-objective evolutionary feature selection for online sales forecasting. Neurocomputing. 2017;234:75–92. doi: 10.1016/j.neucom.2016.12.045. [DOI] [Google Scholar]
- 35.Deb K., Pratab A., Agarwal S., Meyarivan T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002;6:182–197. doi: 10.1109/4235.996017. [DOI] [Google Scholar]
- 36.Bao C., Xu L., Goodman E.D., Cao L. A novel non-dominated sorting algorithm for evolutionary multi-objective optimization. J. Comput. Sci. 2017;23:31–43. doi: 10.1016/j.jocs.2017.09.015. [DOI] [Google Scholar]
- 37.Jiménez F., Sánchez G., Vasant P. A Multi-objective Evolutionary Approach for Fuzzy Optimization in Production Planning. J. Intell. Fuzzy Syst. 2013;25:441–455. [Google Scholar]
- 38.Deb K., Jain H. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints. IEEE Trans. Evol. Comput. 2014;18:577–601. doi: 10.1109/TEVC.2013.2281535. [DOI] [Google Scholar]
- 39.Frank E., Witten I.H. Generating Accurate Rule Sets without Global Optimization. Department of Computer Science, University of Waikato; Waikato, New Zealand: 1998. pp. 144–151. [Google Scholar]
- 40.Witten I.H., Frank E., Hall M.A. Introduction to Weka. In: Witten I.H., Frank E., Hall M.A., editors. Data Mining: Practical Machine Learning Tools and Techniques. 3rd ed. Morgan Kaufmann; Boston, MA, USA: 2011. pp. 403–406. (The Morgan Kaufmann Series in Data Management Systems). [Google Scholar]
- 41.Michalski R.S. On the quasi-minimal solution of the general covering problem; Proceedings of the V International Symposium on Information Processing (FCIP 69); Bled, Yugoslavia. 8–11 October 1969; pp. 125–128. [Google Scholar]
- 42.Quinlan J.R. C4. 5: Programs for Machine Learning. Elsevier; Amsterdam, The Netherlands: 2014. [Google Scholar]
- 43.Rajput A., Aharwal R.P., Dubey M., Saxena S., Raghuvanshi M. J48 and JRIP rules for e-governance data. IJCSS. 2011;5:201. [Google Scholar]
- 44.Cohen W.W. Fast effective rule induction; Proceedings of the Twelfth International Conference on Machine Learning; Tahoe City, CA, USA. 9–12 July 1995; pp. 115–123. [Google Scholar]
- 45.Fürnkranz J., Widmer G. Incremental reduced error pruning; Proceedings of the Eleventh International Conference; New Brunswick, NJ, USA. 10–13 July 1994; pp. 70–77. [Google Scholar]
- 46.Holte R.C. Very simple classification rules perform well on most commonly used datasets. Mach. Learn. 1993;11:63–90. doi: 10.1023/A:1022631118932. [DOI] [Google Scholar]
- 47.Mukhopadhyay A., Maulik U., Bandyopadhyay S., Coello C.A.C. A Survey of Multiobjective Evolutionary Algorithms for Data Mining: Part I. IEEE Trans. Evol. Comput. 2014;18:4–19. doi: 10.1109/TEVC.2013.2290086. [DOI] [Google Scholar]
- 48.Mukhopadhyay A., Maulik U., Bandyopadhyay S., Coello C.A.C. Survey of Multiobjective Evolutionary Algorithms for Data Mining: Part II. IEEE Trans. Evol. Comput. 2014;18:20–35. doi: 10.1109/TEVC.2013.2290082. [DOI] [Google Scholar]
- 49.Ishibuchi H., Murata T., Turksen I. Single-objective and two-objective genetic algorithms for selecting linguistic rules for pattern classification problems. Fuzzy Sets Syst. 1997;89:135–150. doi: 10.1016/S0165-0114(96)00098-X. [DOI] [Google Scholar]
- 50.Srinivas M., Patnaik L. Adaptive probabilities of crossover and mutation in genetic algorithms. IEEE Trans. Syst. Man Cybern B Cybern. 1994;24:656–667. doi: 10.1109/21.286385. [DOI] [Google Scholar]
- 51.Zwitter M., Soklic M. Breast Cancer Data Set. Yugoslavia. [(accessed on 5 September 2018)]; Available online: http://archive.ics.uci.edu/ml/datasets/Breast+Cancer.
- 52.Thrun S. MONK’s Problem 2 Data Set. [(accessed on 5 September 2018)]; Available online: https://www.openml.org/d/334.
- 53.Thrun S.B., Bala J., Bloedorn E., Bratko I., Cestnik B., Cheng J., Jong K.D., Dzeroski S., Fahlman S.E., Fisher D., et al. The MONK’s Problems A Performance Comparison of Different LearningAlgorithms. [(accessed on 5 September 2018)]; Available online: http://digilib.gmu.edu/jspui/bitstream/handle/1920/1685/91-46.pdf?sequence=1.
- 54.Metz C.E. Basic principles of ROC analysis. Semin. Nucl. Med. 1978;8:283–298. doi: 10.1016/S0001-2998(78)80014-2. [DOI] [PubMed] [Google Scholar]
- 55.Fawcett T. An Introduction to ROC Analysis. Pattern Recogn. Lett. 2006;27:861–874. doi: 10.1016/j.patrec.2005.10.010. [DOI] [Google Scholar]
- 56.Hand D.J. Measuring classifier performance: A coherent alternative to the area under the ROC curve. Mach. Learn. 2009;77:103–123. doi: 10.1007/s10994-009-5119-5. [DOI] [Google Scholar]
- 57.Zitzler E., Deb K., Thiele L. Comparison of multiobjective evolutionary algorithms: Empirical results. Evol. Comput. 2000;8:173–195. doi: 10.1162/106365600568202. [DOI] [PubMed] [Google Scholar]
- 58.Zitzler E., Thiele L., Laumanns M., Fonseca C., Grunert da Fonseca V. Performance Assessment of Multiobjective Optimizers: An Analysis and Review. IEEE Trans. Evol. Comput. 2002;7:117–132. doi: 10.1109/TEVC.2003.810758. [DOI] [Google Scholar]
- 59.Kohavi R. A study of cross-validation and bootstrap for accuracy estimation and model selection; Proceedings of the 14th International Joint Conference on Artificial Intelligence (II); Montreal, QC, Canada. 20–25 August 1995; pp. 1137–1143. [Google Scholar]
- 60.Jiménez F., Jodár R., Sánchez G., Martín M., Sciavicco G. Multi-Objective Evolutionary Computation Based Feature Selection Applied to Behaviour Assessment of Children; Proceedings of the 9th International Conference on Educational Data Mining EDM 2016; Raleigh, NC, USA. 29 June–2 July 2016; pp. 1888–1897. [Google Scholar]