Abstract
By comparing attributes of objects in an information system, the advantage matrix on the object set is established in this paper. The contributions can be identified as follows: (1) The advantage degree is proposed by the accumulation of the advantage matrix. (2) Based on the advantage matrix, the advantage (disadvantage) neighborhood approximation operator and the advantage (disadvantage) correlation approximation operator are defined and studied. Based on these two new operators, the neighborhood degree and the correlation degree are presented. The relationships between them are also investigated to demonstrate the value of the proposed method. (3) Finally, based on the above three degrees, new algorithms are designed, in which the effectiveness and robustness of the algorithms are analyzed by practical examples.
Keywords: Advantage matrix, Advantage (disadvantage) neighborhood approximation operator, Advantage (disadvantage) correlation approximation operator, Advantage (neighborhood, correlation), Decision-making
Introduction
Many decision-making situations in real life need to consider several criteria (Liang et al. 2015; Lin et al. 2013; Pearman 2014; Qian et al. 2014; Sun et al. 2017, 2017). Multiple-attribute decision making (MADM) is an important branch of decision-making. MADM methods attempt to select the best alternative(s) from a set of alternatives with respect to several criteria. Some of the classic methods of MADM have been proposed, such as ELECTRE method (Figueira et al. 2009) and PROMETHEE method (Brans and Vincke 1986) which are based on “Outranking Relationship", AHP method constructed with the concept of Hierarchy (Belton 1986; Tian et al. 2018), MAVT method based on “Attribute Value function" (Ferretti et al. 2014) and TOPSIS method (Hwang and Yoon 1981; Fan et al. 2013) and EDAS (Evaluation Based On Distance from Average Solution) method (Ghorabaee et al. 2015; Canciglieri et al. 2015). These methods with clear concept and strong practicability, are known as effective ranking approaches and have been widely used in the research of various MADM problems. For example: Dragan and Fatih (Pamučar and Ecer 2020) propose a novel subjective weighting method called the Fuzzy Full Consistency Method (FUCOM-F) for determining weights as accurately as possible under fuzziness, and apply this method to the green supplier evaluation problem. In order to meet the challenge of the COVID-19 (COronaVIrus Disease-2019) pandemic, health systems must adjust to new circumstances and establish separate hospitals exclusive for patients infected with SARS-CoV-2 virus. Žižović et al. (2021) put forward a multiple-criteria model for the evaluation and selection of nurses for COVID-19 hospitals and so on.
Rough set theory (RST) (Pawlak 1982) is a mathematical approach to imprecision, vagueness and uncertainty in data analysis. In RST, lower and upper approximations were defined to characterize a concept, namely, a subset of the universe. The main advantage of RST in data analysis is that it does not need any preliminary or additional information about data. RST has been widely applied to many fields, e.g., forecasting (Yu et al. 2020, 2019, 2019), medical diagnosis (Pattaraintakorn and Cercone 2008), machine learning (Hong et al. 2008, 2002), decision-making (Salamó and López-Sánchez 2011; Son et al. 2012; Swiniarski and Skowron 2003; Tian et al. 2011; Xiong et al. 2012; Zeleny 1976; Ye et al. 2021; Wang et al. 2021), pattern recognition (Wang and Wang 2009), case-based reasoning (Huang and Tseng 2004), and data mining (Lingras and Yao 1998; Yamaguchi 2009).
RST (Pawlak 1997; Pawlak and Slowinski 1994) is a valid mathematical approach to deal with the decision analysis, and has attracted much of attentions. In the past few years, many rough set-based MADM methods were presented. Based on covering (I, T)-fuzzy rough sets, Zhang et al. (2019) defined four kinds of fuzzy -covering models, and designed a fuzzy rough TOPSIS method. By triangular norms and fuzzy implication operators, Jiang et al. (2018) defined four types of fuzzy -covering variable precision (I, T)-fuzzy rough set models, and designed another kind of fuzzy rough TOPSIS method. In the intuitionistic fuzzy environment, Zhang et al. (2019) defined covering (I, T)-intuitionistic fuzzy rough set model based on D’eer fuzzy domain operator, and designed intuitionistic fuzzy rough TOPSIS method. At the same time, from the perspective of classical indiscernibility relation and big data, Yu et al. (2019) gave the attribute oriented -indiscernibility relation, then constructed the -rough set model, and finally successfully integrate the -rough set model and TOPSIS method to establish a kind of rough TOPSIS method, which is applied to student performance evaluation. Yu et al. (2019) used the -indiscernibility relation to construct an -rough fuzzy set model from the perspective of granular computing. Then they successfully integrated the -rough fuzzy set model with PROMETHEE, and established a class of rough fuzzy PROMETHEE method, and applied it to the analysis of enterprise development. Based on the -rough fuzzy set model, Yu et al. (2020) designed a multi-attribute decision-making method (the prediction method based on the minimum deviation), and discussed the application of this method.
Based on the review conducted above, some genuine challenges are identified which are presented in a nutshell below:
In traditional methods, dimension and attribute weight are used as a priori information, and different processing methods lead to great differences in decision-making results. Next, the TOPSIS method is illustrated as an example. In Table 1, the TOPSIS based on 3 determining attribute weight methods (Entropy method (Zou et al. 2006), Coefficient of variation method (Faber and Korn 1991), Cumulation method) and 4 normalization methods (Max method, Sum method, Max-Min method, Vector method) carries out the decision analysis, and the decision results are shown in Fig. 1a, b. In Fig. 1a, the influence of weight on sorting results is analyzed, the distance is European distance, and the vector method is used for normalization. In Fig. 1b, the influence of normalization on sorting results is analyzed, the distance is European distance, and all attribute weights are set to the same.
Table 1.
An information system
| S | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| 0.1341 | 0.4417 | 0.4561 | 0.2982 | 0.0809 | 0.3381 | 0.6035 | 0.6925 | 0.6079 | |
| 0.2126 | 0.0133 | 0.1017 | 0.0464 | 0.7772 | 0.2940 | 0.5261 | 0.5567 | 0.7413 | |
| 0.8949 | 0.8972 | 0.9954 | 0.5054 | 0.9051 | 0.7463 | 0.7297 | 0.3965 | 0.1048 | |
| 0.0715 | 0.1967 | 0.3321 | 0.7614 | 0.5338 | 0.0103 | 0.7073 | 0.0616 | 0.1279 | |
| 0.2425 | 0.0934 | 0.2973 | 0.6311 | 0.1092 | 0.0484 | 0.7814 | 0.7802 | 0.5495 | |
| 0.0538 | 0.3074 | 0.0620 | 0.0899 | 0.8258 | 0.6679 | 0.2880 | 0.3376 | 0.4852 |
Fig. 1.
Sorting results based on TOPSIS
-
2.
In the process of decision method design, the distance (Euclidean distance, City block metric, Chebyshev distance, Mahalanobis distance, etc.) is usually used to measure the quality of objects. Choosing different distances also affects the decision results. Next, the TOPSIS method is illustrated as an example, too. In Table 1, the TOPSIS based on 4 distance carries out the decision analysis, and the decision results are shown in Fig. 1c. In terms of decision semantics, although weight is used to adjust the importance of attributes in the distance, it do not think about the internal influence between attributes, such as incompatibility and non operability.
-
3.
In the existing methods, the advantage relations between objects are established from various angles, but these relations often have nonlinear relationships, which is difficult to meet the needs of optimal ranking.
Motivated by these challenges, the major contributions of this research are threefold:
Through the comparison between attributes one by one, the advantage matrix is established to intuitively judge the advantage attributes among objects. From the perspective of decision semantics, it only compares the same attributes of different objects without considering the impact of attributes. The decision-making method based on advantage matrix design avoids dimensional processing, so as to weaken the influence of prior information on decision-making results.
The approximation operator is the core of domain rough set, and domain rough set does not need any prior information other than the data set required by the problem when describing or processing the problem, so the description or processing of the problem is more objective. The decision-making method based on approximation operators is designed to avoid the determination of attribute weight, so as to weaken the influence of a prior information on the decision-making results.
The non-linear advantage relation between objects is constructed based on the advantage matrix, the approximation operator of the advantage relation is established, and the neighborhood degree and the correlation degree are proposed to global analysis, further analysis of the object sequencing, so as to achieve the optimal ranking of objects.
The rest of this paper is organized as follows. In Sect. 2, based on the advantage matrix of objects, the neighborhood approximation operators and correlation approximation operators are constructed, together with investigating the properties of four types of approximations. In Sect. 3, two new MADM methods are designed according to the advantage matrix of objects. The experiments demonstrate that the synthetical results are effective and robust.
Main results
An information system (or knowledge representation system) is a finite table. The rows are labeled by objects, the columns are labeled by attributes, and the entries of the table are in their attribute-values. An information system is a 4-tuple , where (1) U is a non-empty finite set of objects; (2) Q is a non-empty finite set of attributes; (3) where is the set of values of the attribute q; (4) is a map such that for every and . For our purpose, throughout this study, let . The cardinality of a finite set X in this study is written as .
Definition 2.1
(Advantage matrix) Given an information system S, its advantage matrix is a matrix, in which the element for an object pair is defined by
| 1 |
The physical meaning of the matrix element is that the object is superior to with respect to any attribute in . An advantage matrix D is nonsymmetric, i.e., , . Therefore, it is sufficient to consider the lower triangle and the upper triangle of the matrix.
Example 2.2
Let be an information system given in Table 2.
Table 2.
An information system
| S | ||||
|---|---|---|---|---|
| 0.10 | 0.20 | 0.10 | 0.40 | |
| 0.70 | 0.25 | 0.45 | 0.70 | |
| 0.45 | 0.15 | 0.65 | 0.85 | |
| 0.30 | 0.50 | 0.40 | 0.90 | |
| 0.55 | 0.60 | 0.70 | 0.25 |
The advantage matrix D of Table 2 is shown below. For the object pair , the entry indicates that the object is superior to by the attribute , and .
Definition 2.3
For an information system S and , if or , then x is completely superior to y, and denoted by or ; if , then x is relatively superior to y, and denoted by or .
Examples are given below to elaborate Definition 2.3.
Example 2.4
(Continuation of Examples 2.2) By Definition 2.3, , , , , , and .
Definition 2.5
For an information system S. Then
If , , , then S is called a complementary information system (CIS).
If , , , then S is called a non-complementary information system (NCIS).
If , , , then S is called a linear-comparable information system (LCIS).
Theorem 2.6
For an information system S and . Then
If S is a CIS, then .
S is a CIS ,
S is an NCIS ,
S is an LCIS , or .
Proof
Proposition 2.7
For an information system S and . Then
and .
. In particular, if , then , , x is called a maximum object.
. In particular, if , then , , x is called a minimum object.
Proof
Definition 2.8
(Advantage degree) The advantage degree (AD) of an object x is defined by
| 2 |
.
Note
In an information system, the advantage degree is used to describe by the advantage attribute of one object to other objects. By comparing the advantage degrees of each object, the advantages and disadvantages of the object are analyzed. This description method not only compares each object (by ), but also compares all objects comprehensively (by ). In particular, , .
In Table 2, based on the total utility value (), , , , , and . The results allow us to believe that the objects , , and are better than , and that , and are not comparable. To achieve a more efficient sorting result in Table 2, the advantage degrees based on the advantage matrix are as follows: , , , , and . Immediately, is superior to ; is superior to ; is superior to ; is superior to .
For the above two methods, incomparable situations will appear for two objects. In order to solve this kind of problem, two kinds of approximate operators on the basis of the advantage matrix are proposed.
Definition 2.9
Let be an information system and .
- The advantage and disadvantage neighborhood approximation operators are respectively defined as:
and3 4 - The advantage and disadvantage correlation approximation operators are respectively defined as:
and5 6
Examples are given below to elaborate Definition 2.9.
Example 2.10
(Continuation of Examples 2.2) By Definition 2.9, we have , , , , , , , , , , , , , , , , , , , .
Note
For an information system, the advantages and disadvantages of the objects are analyzed by combining the advantage and disadvantage neighborhood. They reflect the advantages and disadvantages of objects in information systems from different perspectives. Assuming that we have six objects , and , in which is superior to ; is superior to ; is superior to ; is superior to ; and is superior to . Then and . By contrast, if and , then is superior to ; is superior to ; is superior to ; is superior to ; and is superior to . The less objects in the advantage neighborhood of the object x, the result in the higher priority x achieves. The more objects in the disadvantage neighborhood of the object x, the result in the higher priority x achieves.
In an information system, represents the set of objects that have an advantage attribute for the object x, represents the set of objects that have a disadvantage attribute for the object x. They explain the correlation between objects in the information system from different views. The semantic expression is as follows: The less objects in the advantage correlation of the object x, the result in the higher priority x achieves. The more objects in the disadvantage correlation of the object x, the result in the higher priority x achieves.
According to the semantic expression of neighborhood and correlation approximation operators, we propose the following two new degrees.
Definition 2.11
(Neighborhood degree and correlation degree) Let be an information system and .
- The neighborhood degree (ND) of an object x is defined as
7 - The correlation degree (CD) of an object x is defined as
8
Proposition 2.12
Let be an information system and . Then
If x is a maximum object, then , , , and .
If x is a minimum object, then , , , and .
If x is a maximum object, then , .
If x is a minimum object, then , .
Proof
Theorem 2.13
Let be an information system and . Then
and .
and .
and .
Proof
According to Definition 2.9, obviously.
If , then . Since , we have , that is, . So, , thus, . Similarly, .
According to (1), (3) is straightforward.
Theorem 2.14
Let be a CIS and . Then
and .
and .
.
.
Proof
Since S is a CIS, we have by Theorem 2.6. By Definition 2.9 (1), we have , so . Thus, . Similarly, we can proof that .
According to Definition 2.1, , . By Definition 2.9, we have . Then . Conversely, suppose , . Then, and . If , then . Since S is a CIS, we have , . That is, , which contradicts . Thus, , . In summary, . Similarly, .
According to (2) and Theorem 2.13 (2), we have . Then, .
It is similar to the proof of (3).
Application
Algorithms and examples
In the sequel, we present the general problem with the best selection from m objects (, ). Here we evaluate and compare with the other objects on the basis of attribute (, ) whose values are known to us. Given a decision information system , where , , and , . follows.
Algorithm 3.1
(Multi attribute decision making based on the advantage matrix)
Input: An information system
Output: A ranking result of all objects.
Step 1: Compute the advantage matrix , the object , according to Definition 2.1;
Step 2: Compute the advantage degree , , according to Definition 2.8;
Step 3: Rank the objects according to the of each object. The bigger the is, the better the object will be,
, or , ;
Step 4: End.
Algorithm 3.2
(Multi attribute decision making based on the neighborhood approximation operator and the correlation approximation operator)
Input: A information system and an increase .
Output: A ranking result of all objects.
Step 1: Compute the advantage matrix , the object , according to Definition 2.1;
Step 2: Compute the neighborhood degree , , according to Definition 2.11 (1);
Step 3: Compute the correlation degree , , according to Definition 2.11 (2);
Step 4: If , and , then ; If , and , then ; If , , then and , where and are mixed to establish ;
Step 5: Rank the objects according to the of each object. The bigger the is, the better the object will be,
, ;
Step 6: End.
To understand Algorithms 3.1 and 3.2 better, they are described by a flowchart as Fig. 2.
Fig. 2.
Next, an example is introduced to demonstrate the two algorithms.
Example 3.3
Let be an information system given in Table 3.
Table 3.
An information system
| S | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 0.1080 | 0.8722 | 0.4519 | 0.4272 | 0.8291 | 0.8776 | 0.5811 | 0.5289 | 0.4445 | 0.3453 | |
| 0.4599 | 0.0522 | 0.3334 | 0.1687 | 0.6266 | 0.0144 | 0.6372 | 0.6944 | 0.0854 | 0.9468 | |
| 0.4509 | 0.2197 | 0.0591 | 0.7517 | 0.5387 | 0.2943 | 0.6513 | 0.2124 | 0.0573 | 0.5202 | |
| 0.5511 | 0.4596 | 0.7409 | 0.3684 | 0.6505 | 0.1799 | 0.8646 | 0.5433 | 0.6295 | 0.9538 | |
| 0.8054 | 0.9585 | 0.5068 | 0.9418 | 0.7266 | 0.9263 | 0.0560 | 0.7025 | 0.7962 | 0.0736 | |
| 0.7009 | 0.7900 | 0.1999 | 0.0172 | 0.0945 | 0.0682 | 0.8169 | 0.9564 | 0.6912 | 0.2070 |
(1) Based on Algorithm 3.1,
- Step 1:
- Compute the advantage matrix .
- Step 2:
Compute the advantage degree, , , , , , and .
- Step 3:
Rank the objects according to the of each object, .
- Step 4:
End.
(2) Based on Algorithm 3.2,
- Step 1:
Compute the advantage matrix D, it is the same as (1).
- Step 2:
Compute the neighborhood degree, , , , , , and .
- Step 3:
According to Step 2, skip.
- Step 4:
Skip.
- Step 5:
Rank the objects according to the ,
- Step 6:
End.
Case analysis
For the sake of proving the effectiveness and robustness of our new algorithms, Student Performance Data Set Student performance (2014) is selected from the UC Irvine Machine Learning Repository (UCI). In this data approach student achievement in secondary education of two Portuguese schools. The data attributes include student grades, demographic, social and school related features) and it was collected by using school reports and questionnaires, as shown in Tables 4 and 5 . Two data sets are provided regarding the performance in two distinct subjects: Mathematics (Mat) and Portuguese language (Por). In Cortez and Silva (2008), the two data sets are modeled under 5-level classification and regression tasks. Important note: the target attribute has a strong correlation with attributes and .
Table 4.
Attribute information
| Attribute | Description (Domain) |
|---|---|
| school | Student’s school (binary: Gabriel Pereira or Mousinho da Silveira) |
| sex | Student’s sex (binary: female or male) |
| age | Student’s age (numeric: from 15 to 22) |
| address | Student’s home address type (binary: 1 - urban or 0 - rural) |
| famsize | Family size (binary: less or equal to 3 or greater than 3) |
| Pstatus | Parent’s cohabitation status (binary: living together or apart) |
| Medu |
Mother’s education (numeric: 0 - none, 0.2 - primary education(4th grade), 0.4 - 5th to 9th grade, 0.6 - secondary education or 0.8 - higher education) |
| Fedu |
Father’s education (numeric: 0 - none, 0.2 - primary education(4th grade), 0.4 - 5th to 9th grade, 0.6 - secondary education or 0.8 - higher education) |
| Mjob | Mother’s job (nominal: ’teacher’, ’health’ care related, civil ’services’ ) |
| Fjob | Father’s job (nominal: ’teacher’, ’health’ care related, civil ’services’) |
| reason |
Reason to choose this school (nominal: close to ’home’, school ’reputation’, ’course’ preference or ’other’) |
| guardian | Student’s guardian (nominal: ’mother’, ’father’ or ’other’) |
| traveltime |
Home to school travel time (numeric: 0.8 - less than 15 min., 0.6 - 15 to 30 min., 0.39 - 30 min. to 1 hour, or 0.19 - greater than 1 hour) |
| studytime |
Weekly study time (numeric: 0.2 - less than 2 hours, 0.4 - 2 to 5 hours, 0.6 - 5 to 10 hours, or 0.8 - greater than 10 hours) |
| failures | Number of past class failures (numeric: n, if , else 4) |
| schoolsup | Extra educational support (binary: 1 - yes or 0 - no) |
| famsup | Family educational support (binary: 1 - yes or 0 - no) |
| paid |
Extra paid classes within the course subject (Math or Portuguese) (binary: 1-yes or 0-no) |
| activities | Extra-curricular activities (binary: yes or no) |
| nursery | Attended nursery school (binary: 1 - yes or 1 - no) |
| higher | Wants to take higher education (binary: 1 - yes or 0 - no) |
| internet | Internet access at home (binary: yes or no) |
| romantic | With a romantic relationship (binary: yes or no) |
| famrel |
Quality of family relationships (numeric: 0.2 - very bad, 0.4 - bad, 0.6 - good, 0.8 - very good or 1 - excellent) |
| freetime |
Free time after school (numeric: 0.2 - very low, 0.4 - low, 0.6 - general, 0.8 - high, 1 - very high) |
| goout | Going out with friends (numeric: from 1 - very low to 5 - very high) |
| Dalc |
Workday alcohol consumption (numeric: 0.9 - very low, 0.6 - low, 0.4 - general, 0.2 - high, 0 - very high) |
| Walc |
Weekend alcohol consumption (numeric: 0.9 - very low, 0.6 - low, 0.4 - general, 0.2 - high, 0 - very high) |
| health |
Current health status (numeric: 0.2 - very bad, 0.4 - bad, 0.6-good, 0.8 - very good or 1 - excellent) |
| attendance | (Rate of) attendance (numeric: , n number of school attendances) |
Table 5.
Decision attribute information
| Target attribute | Description (Domain) |
|---|---|
|
First period grade (numeric: 5 - very good (16-20), 4 - good (14-15), 3 - satisfactory (12-13), 2 - sufficient (10-11) , 1 - fail (0-9) |
|
|
Second period grade (numeric: 5 - very good (16-20), 4 - good (14-15), 3 - satisfactory (12-13), 2 - sufficient (10-11) , 1 - fail (0-9) |
|
|
Final grade (numeric: 5 - very good (16-20), 4 - good (14-15), 3 - satisfactory (12-13), 2 - sufficient (10-11) , 1 - fail (0-9) |
In the target item, the higher the level of the object, the better the object is
It is generally believed that , and are related to students’ grades. Based on Student Performance Data Set Student performance (2014), 4 new information systems are set up, as shown in Tables 6, 7, 8 and 9. In Tables 7 and 9 , and represent the grades of 3 stages and actual results, as shown. Tables 7 and 9 are used as comparison indicators to verify whether the sorting results are correct.
Table 6.
An information system M based on Mathematics course
| M | . | . | . | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 1.00 | 0.80 | 0.80 | 0.60 | . | 0 | 0 | . | 0.80 | 0.60 | . | 0.60 | 0.94 | |
| 1.00 | 0.20 | 0.20 | 0.80 | . | 1.00 | 0 | . | 1.00 | 0.60 | . | 0.60 | 0.96 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 1.00 | 0.80 | 0.60 | 0.80 | . | 1.00 | 1.00 | . | 1.00 | 0.60 | . | 0.60 | 1.00 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 1.00 | 0.80 | 0.80 | 0.80 | . | 1.00 | 1.00 | . | 0.80 | 1.00 | . | 0.60 | 1.00 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 1.00 | 0.20 | 0.20 | 0.80 | . | 0 | 0 | . | 0.60 | 0.40 | . | 1.00 | 0.95 |
Table 7.
Three period achievements based on Mathematics course
| M | |||
|---|---|---|---|
| 1 | 1 | 1 | |
| 1 | 1 | 1 | |
| . | . | . | . |
| 1 | 1 | 1 | |
| . | . | . | . |
| 1 | 1 | 2 | |
| . | . | . | . |
| 1 | 1 | 1 |
Table 8.
An information system P based on Portuguese language course
| P | . | . | . | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 1.00 | 0.80 | 0.80 | 0.60 | . | 0 | 0 | . | 0.80 | 0.60 | . | 0.6000 | 0.96 | |
| 1.00 | 0.20 | 0.20 | 0.80 | . | 1.00 | 0 | . | 1.00 | 0.60 | . | 0.60 | 0.98 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 1.00 | 0.60 | 0.60 | 0.80 | . | 1.00 | 0 | . | 0.80 | 0.60 | . | 0.8000 | 0.94 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 1.00 | 0.60 | 0.40 | 0.80 | . | 0 | 0 | . | 1.00 | 0.8000 | . | 0.2000 | 0.96 | |
| . | . | . | . | . | . | . | . | . | . | . | . | . | . |
| 0 | 0.60 | 0.40 | 0.39 | . | 0 | 0 | . | 0.80 | 0.80 | . | 1.00 | 0.96 |
Table 9.
Three period achievements based on Portuguese language course
| P | |||
|---|---|---|---|
| 1 | 2 | 2 | |
| 1 | 2 | 2 | |
| . | . | . | . |
| 2 | 3 | 4 | |
| . | . | . | . |
| 4 | 5 | 5 | |
| . | . | . | . |
| 2 | 2 | 2 |
Results and discussions
Sorting results and Spearman rank correlation coefficient
To demonstrate the application and effect of Algorithms 3.1 and 3.2 , these two new algorithms are used to deal with two data sets (Mat and Por), . To show the experimental results, 3 object sets are randomly selected from Tables 6 and 8 , respectively. are selected from Table 6. , , are selected from Table 8. Algorithms 3.1 and 3.2 are use to analyze the above 6 object sets, and the results are shown in Figs. 3 and 4.
Fig. 3.
The sort based on Mathematics
Fig. 4.
The sort based on Portuguese
In Fig. 3a, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 7. In Fig. 3b, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 7. In Fig. 3c, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 7.
In Fig. 4a, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 9. In Fig. 4b, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 9. In Fig. 4c, Algorithms 3.1 and 3.2 are used to sort , and compare sorting results with Table 9.
To better describe the consistency between the sorting results and the actual results, the Spearman rank correlation coefficient or Spearman (Myers et al. 2013) is used to analyze the results in Figs. 3 and 4 , the results are shown in Tables 10, 11, 12, 13, 14 and 15. It is widely known that if the Spearman between sorting results of two methods belongs to [0, 1], then the two methods are positively correlated. Moreover, as a commonly accepted notion, if the Spearman between sorting results of two methods is greater than 0.6, then the correlation degree between the two methods is high.
Table 10.
Spearman based on Fig. 3a
| Alg.3.2 | Alg.3.1 | Mat | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.9 | 0.9 |
| Alg.3.1 | 0.9 | 1 | 0.7 |
| Mat | 0.9 | 0.7 | 1 |
Table 11.
Spearman based on Fig. 3b
| Alg.3.2 | Alg.3.1 | Mat | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.9 | 0.9 |
| Alg.3.1 | 0.9 | 1 | 0.8 |
| Mat | 0.9 | 0.8 | 1 |
Table 12.
Spearman based on Fig. 3c
| Alg.3.2 | Alg.3.1 | Mat | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.6 | 0.8 |
| Alg.3.1 | 0.6 | 1 | 0.6 |
| Mat | 0.8 | 0.6 | 1 |
Table 13.
Spearman based on Fig. 4a
| Alg.3.2 | Alg.3.1 | Por | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.9 | 0.9 |
| Alg.3.1 | 0.6 | 1 | 0.7 |
| Por | 0.9 | 0.7 | 1 |
Table 14.
Spearman based on Fig. 4b
| Alg.3.2 | Alg.3.1 | Por | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.6 | 0.9 |
| Alg.3.1 | 0.6 | 1 | 0.7 |
| Por | 0.9 | 0.7 | 1 |
Table 15.
Spearman based on Fig. 4c
| Alg.3.2 | Alg.3.1 | Por | |
|---|---|---|---|
| Alg.3.2 | 1 | 0.9 | 0.9 |
| Alg.3.1 | 0.9 | 1 | 0.8 |
| Por | 0.9 | 0.8 | 1 |
In Figs. 3 and 4 , the broken line trend of each subgraph is basically consistent, which also shows that the results obtained by Algorithms 3.1 and 3.2 are basically consistent with the actual sorting results. In Tables 10, 11, 12, 13, 14 and 15, the Spearman among Algorithms 3.1, 3.2 and the real sorting results are greater than or equal to 0.6. In particular, the most results are greater than or equal to 0.8. In the summary, there is a high correlation among Algorithms 3.1, 3.2 and the real results, and Algorithms 3.1 and 3.2 are effective.
Random results and accuracy
The number of objects of the two data sets (Mat and Por) is relatively large, and the real ranking is 5 levels, so it is unreasonable to sort all objects. This paper uses the random sampling method to verify the accuracy of Algorithms 3.1 and 3.2 . The specific process is as follows:
For Fig. 5a, the specific process is as follows: (1) two objects are randomly selected from 395 objects in the Mathematical data set; (2) the objects are sorted by comparing the NDCD values and Mat (), if the two sorting results are consistent, the result of Algorithm 3.2 is accurate; (3) Repeat (1) and (2) 1000 times, and count the proportion of consistent sorting results as the accuracy of Algorithm 3.2; (4) Repeat (1), (2) and (3) 1000 times to obtain the accuracy of 1000 times, and draw the diagram. Other subgraphs of Fig. 5 are similarly processed. Overall, the accuracies of Algorithms 3.2 and 3.1 are distributed between 0.60 and 0.72, and the whole random process is randomly selected times (Select 2 of the 395 objects, ).
Fig. 5.
Stochastic analysis based on the Mathematic data set
Comparing the process in Fig. 5, the Mathematic data set is replaced by the Portuguese data set. Other processes are the same, and the results are shown in Fig. 6, the accuracies of Algorithms 3.2 and 3.1 are distributed between 0.62 and 0.74, and the whole random process is randomly selected times (Select 2 of the 649 objects, ).
Fig. 6.
Stochastic analysis based on the Portuguese data set
Accuracy comparison and stability analysis
In order to show the accuracy of Algorithms 3.2, 3.1, PROMETHEE, TOPSIS, and SAW, the detailed test process is as follows:
For Fig. 7a, the specific process is as follows: (1) Sort 395 objects in the Mathematical data set by Algorithms 3.2, 3.1, PROMETHEE, TOPSIS, and SAW; (2) Generate all binary combinations, ; (3) Based on all binary combinations, the proportion that the sorting of Algorithm 3.2 (Algorithm 3.1, PROMETHEE, TOPSIS, SAW) is consistent with Mat () is recorded as the accuracy of Algorithm 3.2 (Algorithm 3.1, PROMETHEE, TOPSIS, SAW), and the results are shown in Fig. 7a (Mat ()). Other broken lines in Fig. 7a shall be subject to the same test process. The accuracy analysis based on the Portuguese data set is shown in Fig. 7b.
Fig. 7.
Accuracy comparison
Overall, the accuracy of Algorithms 3.2 and 3.1 is higher than that of PROMETHEE, TOPSIS, and SAW in the mathematical data set; In the Portuguese data set, Algorithm 3.2 is exactly the same as PROMETHEE, TOPSIS, and SAW, but Algorithm 3.1 is lower than PROMETHEE, TOPSIS, and SAW.
For the two data sets (Mat and Por), the sensitivity of 5 algorithms (Algorithms 3.2, 3.1, PROMETHEE, TOPSIS, SAW) is analyzed by deleting attributes. The specific test process is as follows:
For Fig. 8, by deleting the attributes one by one, the influence of attribute change on the accuracy of 5 algorithms (Algorithms 3.2, 3.1, PROMETHEE, TOPSIS, SAW) is verified. Combined with Figs. 7 and 8, the accuracy change is shown in Fig. 9.
Fig. 8.
Accuracy when deleting attributes based on Mat and Por
Fig. 9.
Accuracy changes when deleting attributes based on Mat and Por
For Fig. 9a–c, PROMETHEE and TOPSIS have the worst stability, followed by SAW, and Algorithms 3.1 and 3.2 have the best stability. For Fig. 9 (d),(e),(f), PROMETHEE has the worst stability, and other methods have similar stability. Overall, SAW is linear and more stable, but its scope of application is limited; PROMETHEE and TOPSIS are non-linear and poor stable, but their scope of application are extensive. Algorithms 3.1 and 3.2 are non-linear, the stability of them is similar to SAW, or even better.
In summary, (1) From the experimental results: Algorithms 3.1 and 3.2 are used for student ranking based on two data sets (Mat and Por). The sorting results of Algorithms 3.1 and 3.2 are compared with the real sorting results. Algorithms 3.1 and 3.2 are highly correlated with the actual results; By comparing Algorithms 3.1 and 3.2 with PROMETHEE, TOPSIS, and SAW, the accurate pair of display Algorithms 3.1 and 3.2 are higher than or equal to other methods; Through random process and deleting attributes, it shows that Algorithms 3.1 and 3.2 are robust and better than other methods. (2) For management implications: On the one hand, in the process of decision analysis, Algorithms 3.1 and 3.2 not only analyze the relationship between objects, but also analyze the impact of all objects on each object. In the process of Algorithm 3.2 design, a new rough set model is introduced. This model does not simply use the rough set model to analyze the decision results, but uses the principle of rough set and combines with the purpose of decision analysis to establish a new rough model with management semantics; On the other hand, most of the existing decision-making methods rely on subjective experience or prior information, but different subjective experience or prior information leads to great differences in decision-making results. Algorithms 3.1 and 3.2 weaken the dependence on subjective experience and prior information, making the decision-making results relatively objective.
Conclusions
The present study is designed to design two multi-attribute decision-making methods based on less subjective experience or prior information. In this study, two new methods based on the advantage matrix are proposed to handle MADM problems. Compared with the real decision results and other methods, the accuracy and robustness of two new multi-attribute decision-making methods are analyzed. For the two new multi-attribute decision-making methods, the two methods rely on all objects, so the accuracy is high. At the same time, the change of a few attributes does not affect the decision result, that is, strong robustness. A number of limitations to our study should be noted, including that the element diversity of advantage matrix is needed. In other words, in Algorithm 3.1, the more diverse the modules of the elements of the advantage matrix, the better the decision-making effect; In Algorithm 3.2, the disadvantage and advantage neighborhood (correlation) approximations of each object are different. On the contrary, if a large number of empty or complete sets appear in the advantage matrix, Algorithms 3.1 and 3.2 fail. Our work clearly has some limitations. Despite this we believe our work could be a new perspective for decision analysis. For specific applications, the two methods are non-linear algorithms, which have a wider scope of application. Specifically, as long as an information system is established (see the previous information system description), and the attribute information is comparable, no other subjective experience or a prior information is required, and finally the decision result is obtained.
As for the directions of future research, we will investigate complementary issues as follows.
According to the characteristics of the advantage matrix, We will extend the advantage matrix to the intuitionistic fuzzy environment, linguistic environment, and the Hesitant fuzzy environment and generalize our methods in such environments.
According to the characteristics of the advantage matrix, the object is analyzed only from the perspective of the advantages of the attributes, and the disadvantages of the attributes are missing. We will establish the advantage-disadvantage matrix and define a new degree based on the advantage-disadvantage matrix for decision analysis.
For an order information system (Greco et al. 2002), the advantage matrix and directed graph between objects will be established, and the Graph Neural Networks (Wu et al. 2020) will be used to sort and cluster the objects.
According to Definition 2.9, the neighborhood operator is aimed at individual objects and is used for individual decision analysis of individual objects. We will define the neighborhood operator of the object group and use it for group decision analysis, such as comparison among multiple sales teams, asset oriented portfolio investment, etc.
Acknowledgements
This work is supported by the Hunan Provincial Natural Science Foundation of China (Nos. 2019JJ50490, 2020JJ5346), and Scientific Research Foundation of Hunan Province Education Department (No. 18C0455).
Declarations
Conflict of interest
The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Bin Yu, Email: yu7bin@hotmail.com.
Zeshui Xu, Email: xuzeshui@263.net.
References
- Belton V. A comparison of the analytic hierarchy process and a simple multi-attribute value function. Eur J Oper Res. 1986;26(1):7–21. doi: 10.1016/0377-2217(86)90155-4. [DOI] [Google Scholar]
- Brans J, Vincke MB. How to select and how to rank projects: the promethee method. Eur J Oper Res. 1986;24(2):228–238. doi: 10.1016/0377-2217(86)90044-5. [DOI] [Google Scholar]
- Canciglieri O, Jr, et al. Multi-attribute method for prioritization of sustainable prototyping technologies. Clean Technol Environ Policy. 2015;17(5):1–9. doi: 10.1007/s10098-015-0962-5. [DOI] [Google Scholar]
- Cortez P, Silva A. Using data mining to predict secondary school student performance. EUROSIS. 2008;13:5–12. [Google Scholar]
- Faber D, Korn H. Applicability of the coefficient of variation method for analyzing synaptic plasticity. Biophys J. 1991;60(5):1288–1294. doi: 10.1016/S0006-3495(91)82162-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fan ZP, Zhang X, Liu Y, Zhang Y. A method for stochastic multiple attribute decision making based on concepts of ideal and anti-ideal points. Appl Math Comput. 2013;219(24):11438–11450. [Google Scholar]
- Ferretti V, Bottero M, Mondini G. Decision making and cultural heritage: An application of the multi-attribute value theory for the reuse of historical buildings. J Cult Herit. 2014;15(6):644–655. doi: 10.1016/j.culher.2013.12.007. [DOI] [Google Scholar]
- Figueira JR, Greco S, Roy B. Electre methods with interaction between criteria: an extension of the concordance index. Eur J Oper Res. 2009;199(2):478–495. doi: 10.1016/j.ejor.2008.11.025. [DOI] [Google Scholar]
- Ghorabaee M, Zavadskas E, Olfat L, Turskis Z. Multi-criteria inventory classification using a new method of evaluation based on distance from average solution (EDAS) Informatica. 2015;26(3):435–451. doi: 10.15388/Informatica.2015.57. [DOI] [Google Scholar]
- Greco S, Matarazzo B, Slowinski R. Rough approximation by dominance relations. Int J Intell Syst. 2002;17(2):153–171. doi: 10.1002/int.10014. [DOI] [Google Scholar]
- Hong TP, Tseng LH, Wang SL. Learning rules from incomplete training examples by rough sets. Expert Syst Appl. 2002;22(4):285–293. doi: 10.1016/S0957-4174(02)00016-7. [DOI] [Google Scholar]
- Hong TP, Lin CE, Lin JH, Wang SL. Learning cross-level certain and possible rules by rough sets. Expert Syst Appl. 2008;34(3):1698–1706. doi: 10.1016/j.eswa.2007.01.038. [DOI] [Google Scholar]
- https://archive.ics.uci.edu/ml/datasets/Student+Performance. Student performance data set. (2014)
- Huang C, Tseng T. Rough set approach to case-based reasoning application. Expert Syst Appl. 2004;26(3):369–385. doi: 10.1016/j.eswa.2003.09.008. [DOI] [Google Scholar]
- Hwang CL, Yoon K. Multiple attribute decision making. Berlin: Springer; 1981. [Google Scholar]
- Jiang H, Zhan J, Chen D. Covering-based variable precision -fuzzy roughsets with applications to multiattribute decision-making. IEEE Trans Fuzzy Syst. 2018;27(8):1558–1572. doi: 10.1109/TFUZZ.2018.2883023. [DOI] [Google Scholar]
- Liang D, Pedrycz W, Liu D, Hu P. Three-way decisions based on decision-theoretic rough sets under linguistic assessment with the aid of group decision making. Appl Soft Comput. 2015;29:256–269. doi: 10.1016/j.asoc.2015.01.008. [DOI] [Google Scholar]
- Lin TY, Yao YY, Zadeh LA (eds) (2013) Data mining, rough sets and granular computing, vol 95. Physica
- Lingras PJ, Yao Y. Data mining using extensions of the rough set model. J Am Soc Inf Sci Technol. 1998;49(5):415–422. doi: 10.1002/(SICI)1097-4571(19980415)49:5<415::AID-ASI4>3.0.CO;2-Z. [DOI] [Google Scholar]
- Myers J, Well A, Lorch R., Jr . Research design and statistical analysis. Milton Park: Routledge; 2013. [Google Scholar]
- Pamučar D, Ecer F. Prioritizing the weights of the evaluation criteria under fuzziness: the fuzzy full consistency method-FUCOM-F. Facta Univ Ser Mech Eng. 2020;18(3):419–437. [Google Scholar]
- Pattaraintakorn P, Cercone N. Integrating rough set theory and medical applications. Appl Math Lett. 2008;21(4):400–403. doi: 10.1016/j.aml.2007.05.010. [DOI] [Google Scholar]
- Pawlak Z. Rough set. Int J Comput Inf Sci. 1982;11(5):341–356. doi: 10.1007/BF01001956. [DOI] [Google Scholar]
- Pawlak Z. Rough set approach to knowledge-based decision support. Eur J Oper Res. 1997;99(1):48–57. doi: 10.1016/S0377-2217(96)00382-7. [DOI] [Google Scholar]
- Pawlak Z, Slowinski R. Decision analysis using rough sets. Int Trans Oper Res. 1994;1(1):107–114. doi: 10.1111/1475-3995.d01-12. [DOI] [Google Scholar]
- Pearman A. Models and methods in multiple criteria decision making. Amsterdam: Elsevier; 2014. [Google Scholar]
- Qian Y, Zhang H, Sang Y, Liang J. Multigranulation decision-theoretic rough sets. Int J Approx Reason. 2014;55(1):225–237. doi: 10.1016/j.ijar.2013.03.004. [DOI] [Google Scholar]
- Salamó M, López-Sánchez M. Rough set based approaches to feature selection for case-based reasoning classifiers. Pattern Recogn Lett. 2011;32(2):280–292. doi: 10.1016/j.patrec.2010.08.013. [DOI] [Google Scholar]
- Son CS, Kim YN, Kim HS, Park HS, Kim MS. Decision-making model for early diagnosis of congestive heart failure using rough set and decision tree approaches. J Biomed Inform. 2012;45(5):999–1008. doi: 10.1016/j.jbi.2012.04.013. [DOI] [PubMed] [Google Scholar]
- Sun B, Ma W, Qian Y. Multigranulation fuzzy rough set over two universes and its application to decision making. Knowl Based Syst. 2017;123:61–74. doi: 10.1016/j.knosys.2017.01.036. [DOI] [Google Scholar]
- Sun B, Ma W, Xiao X. Three-way group decision making based on multigranulation fuzzy decision-theoretic rough set over two universes. Int J Approx Reason. 2017;81:87–102. doi: 10.1016/j.ijar.2016.11.001. [DOI] [Google Scholar]
- Swiniarski RW, Skowron A. Rough set methods in feature selection and recognition. Pattern Recogn Lett. 2003;24(6):833–849. doi: 10.1016/S0167-8655(02)00196-4. [DOI] [Google Scholar]
- Tian D, Zeng XJ, Keane JA. Core-generating approximate minimum entropy discretization for rough set feature selection in pattern classification. Int J Approx Reason. 2011;52(6):863–880. doi: 10.1016/j.ijar.2011.03.001. [DOI] [Google Scholar]
- Tian G, Zhang H, Feng Y, Wang D, Peng Y, Jia H. Green decoration materials selection under interior environment characteristics: a grey-correlation based hybrid mcdm method. Renew Sustain Energy Rev. 2018;81:682–692. doi: 10.1016/j.rser.2017.08.050. [DOI] [Google Scholar]
- Wang H, Wang S. Discovering patterns of missing data in survey databases: an application of rough sets. Expert Syst Appl. 2009;36(3):6256–6260. doi: 10.1016/j.eswa.2008.07.010. [DOI] [Google Scholar]
- Wang W, Zhan J, Zhang C. Three-way decisions based multi-attribute decision making with probabilistic dominance relations. Inf Sci. 2021;559:75–96. doi: 10.1016/j.ins.2021.01.028. [DOI] [Google Scholar]
- Wu Z, Pan S, Chen F, et al. A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst. 2020;32(1):4–24. doi: 10.1109/TNNLS.2020.2978386. [DOI] [PubMed] [Google Scholar]
- Xiong W, Su Q, Li J. The group decision-making rules based on rough sets on large scale engineering emergency. Syst Eng Procedia. 2012;4:331–337. doi: 10.1016/j.sepro.2011.11.083. [DOI] [Google Scholar]
- Yamaguchi D. Attribute dependency functions considering data efficiency. Int J Approx Reason. 2009;51(1):89–98. doi: 10.1016/j.ijar.2009.08.002. [DOI] [Google Scholar]
- Ye J, Zhan J, Ding W, et al. A novel fuzzy rough set model with fuzzy neighborhood operators. Inf Sci. 2021;544:266–297. doi: 10.1016/j.ins.2020.07.030. [DOI] [Google Scholar]
- Yu B, Cai M, Li Q. A -rough set model and its applications with TOPSIS method to decision making. Knowl Based Syst. 2019;165:420–431. doi: 10.1016/j.knosys.2018.12.013. [DOI] [Google Scholar]
- Yu B, Guo L, Li Q. A characterization of novel rough fuzzy sets of information systems and their application in decision making. Expert Syst Appl. 2019;122:253–261. doi: 10.1016/j.eswa.2019.01.018. [DOI] [Google Scholar]
- Yu B, Cai M, Dai J, Li Q. A novel approach to predictive analysis using attribute-oriented rough fuzzy sets. Expert Syst Appl. 2020;161:113644. doi: 10.1016/j.eswa.2020.113644. [DOI] [Google Scholar]
- Zeleny M (ed) (1976) The theory of the displaced ideal. In: Multiple criteria decision making Kyoto 1975. Lecture Notes in Economics and Mathematical Systems (Operations Research), vol 123. Springer, Berlin, Heidelberg
- Zhang K, Zhan J, Wu W. Fuzzy -covering based -fuzzy rough set models and applications to multi-attribute decision-making. Comput Ind Eng. 2019;128:605–621. doi: 10.1016/j.cie.2019.01.004. [DOI] [Google Scholar]
- Zhang L, Zhan J, Xu Z. Covering-based generalized IF rough sets with applications to multi-attribute decision-making. Inf Sci. 2019;478:275–302. doi: 10.1016/j.ins.2018.11.033. [DOI] [Google Scholar]
- Žižović M, Pamucar D, Miljković B, Karan A. Multiple-criteria evaluation model for medical professionals assigned to temporary SARS-CoV-2 hospitals. Decis Mak Appl Manag Eng. 2021;4(1):153–173. doi: 10.31181/dmame2104153m. [DOI] [Google Scholar]
- Zou ZH, Yi Y, Sun JN. Entropy method for determination of weight of evaluating indicators in fuzzy synthetic evaluation for water quality assessment. J Environ Sci. 2006;18(5):1020–1023. doi: 10.1016/S1001-0742(06)60032-6. [DOI] [PubMed] [Google Scholar]









