Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2019 Mar 3;2019:5370763. doi: 10.1155/2019/5370763

Study on Hesitant Fuzzy Information Measures and Their Clustering Application

Jin-hui Lv 1,, Si-cong Guo 1, Fang-fang Guo 2
PMCID: PMC6421787  PMID: 30944555

Abstract

At present, research on hesitant fuzzy operations and measures is based on equal length processing, and an equal length processing method will inevitably destroy the original data structure and change the data information. This is an urgent problem to be solved in the development of hesitant fuzzy sets. Aiming at solving this problem, this paper firstly defines a hesitant fuzzy entropy function as the measure of the degree of uncertainty of hesitant fuzzy information and then proposes the concept of hesitant fuzzy information feature vector. The hesitant fuzzy distance measure and similarity measure are studied based on the information feature vector. Finally, the hesitant fuzzy network clustering method based on similarity measure is given, and the effectiveness of our algorithm through a numerical example is illustrated.

1. Introduction

Torra and Narukawa [1, 2] extended fuzzy sets [3] to hesitant fuzzy sets (HFSs) because they found that, under a group setting, it is difficult to determine the membership of an element to a set due to doubts between a few different values. For example, two DMs discuss the membership degree of x into A. One wants to assign 0.4 and the other 0.6, and they cannot persuade with each other; thus the membership degrees of x into A can be represented by {0.4, 0.6}. This is obviously different from fuzzy number 0.4 (or 0.6) and the intuitionistic fuzzy number (0.4, 0.6). Therefore, hesitant fuzzy sets can better simulate the hesitant preferences of decision-makers. Since it was put forward, the hesitant fuzzy set has received extensive attention from scholars at home and abroad. The main research work is concentrated in the following aspects: (1) research on various measures in the hesitant fuzzy environment [410]; (2) research on the integration operator of hesitant fuzzy information [1116]; and (3) the expansion of hesitant fuzzy set theory [1722].

It should be pointed out that the present researches on the operation, sorting, and various measures of hesitant fuzzy sets require that the hesitant fuzzy elements have the same length. In practical application, the length of hesitant fuzzy element is different. The method proposed in [2] is adding some elements to a shorter hesitant fuzzy element, making it equal to another hesitant fuzzy element, or repeating their elements in order to obtain two series with the same length [23]. Obviously, these methods will destroy the original data structure and change the data information. How to overcome the shortcomings has become an urgent problem to be solved in the development process of hesitant fuzzy sets.

Clustering is a basic technique, which is often utilized in a primary step of analyzing unlabeled data with the goal of summarizing structural information [24]. In practical applications, the clustering data are mostly uncertain or fuzzy. To solve the problem of data clustering in different fuzzy environments, fuzzy clustering algorithms [25], intuitionistic fuzzy clustering algorithms [26], and 2-type fuzzy clustering algorithms [27] have been proposed. However, in the group of decision-making environment, the decision information is more suitable to express hesitant fuzzy sets, and the algorithm mentioned above is not suitable for handling the clustering problem of this type of information. If the fuzzy logic is used to handle it, generally take the average value of preference information that are provided by experts or can take the minimum range containing all of the preference information, that is, convert the hesitant fuzzy information into interval value information for processing. This method of data processing is bound to change the original preference information that provided by experts; as a result, the research of clustering problem under the hesitant fuzzy information has a certain scientific significance. One of the advantages of applying the hesitant fuzzy set is that clustering hesitant and vague information permits us to find patterns among hesitant fuzzy data. At present, the clustering researches under the hesitant fuzzy environment are still at the its initial stage, and Chen et al. [28] used the correlation coefficient of hesitant fuzzy set to construct hesitant fuzzy relationship matrix and then conducted hesitant fuzzy clustering analysis based on the relation of equivalence. In order to obtain an equivalence relation matrix, a fuzzy relation matrix needs to be iterated continuously, which not only loses information but also has a large amount of calculation [29]. Due to the existence of uncertainty for the similarity measure of samples, leading to the clustering, results were not precise enough and the divided categories were inconsistent with the fact. In [4], the hesitant fuzzy similarity measure formula based on distance was proposed. The measurement is inconsistent with the facts sometimes, and the resolution is not high enough; in the literature [29], a hesitant fuzzy clustering method based on agglomerative hierarchical clustering [30] was proposed. This method needs to use a hesitant fuzzy average operator to calculate the clustering center repeatedly, and the calculation amount is large; in the literature [31], a hesitant fuzzy clustering algorithm based on minimal spanning tree was proposed. The distance of hesitant fuzzy set used in this method is put forward based on the literature [4], which also has the shortcoming of low resolution and sometimes inconsistent with the fact; in the literature [32], from the point of view of information theory, hesitant fuzzy relative entropy and symmetric interactive entropy are proposed, a new kind of hesitant fuzzy similarity degree is proposed, which is combined with the idea of TOPSIS, and a hesitant fuzzy clustering method is proposed based on the traditional netting clustering method. The premise of all the above methods in the measurement and operation is that the data are equal in length, which is not satisfied by the hesitant fuzzy set. Therefore, it is necessary to add artificial elements for equal length processing, and the processed data will inevitably change the original data information and affect the clustering results.

Based on the above analysis, this paper firstly proposes the concept of hesitant fuzzy entropy function and hesitant fuzzy information feature vector, aiming at solving the problem of processing data of hesitant fuzzy set, sorting, and various measures in the study of different lengths. Furthermore, the hesitant fuzzy uncertainty measure, distance measure, and similarity measure are studied. Finally, based on the similarity measure and the traditional network clustering method, the network clustering method for hesitant fuzzy information is given. And then we illustrate its effectiveness via numerical examples.

2. Preliminary

Definition 1 (1, 2). —

Let X={x1,x2,…, xm} be a fixed set; a hesitant fuzzy set (HFS) on X is represented by a function that when applied to X, it returns a subset of [0, 1], which can be expressed by a mathematical symbol:

H=<x,hHx>xX, (1)

where hH(x) is a set of some values in [0,1], denoting the possible membership degrees of the element xX to the set H. hH(x) is called the hesitant fuzzy number or hesitant fuzzy element. If it does not cause confusion, it can be abbreviated as h=hH(x). The hesitant fuzzy number can be expressed in more detail as h=H{γ1, γ2,…, γl}. Among which, l denotes the number of elements in a hesitant fuzzy number h. Obviously, when l=1, the hesitant fuzzy set H degenerates into the traditional fuzzy set.

Definition 2 (1, 2). —

Set X={x1,x2,…, xm} as a given nonempty set, then Hc={<x, hHc(x) > ∣xX} is the complement of the hesitant fuzzy set H, among which

hHcx=γhHx1γ. (2)

Distance measure and similarity measure are important research contents in fuzzy set theory and have a wide application background. In the literature [4], the axiomatic definitions of distance and similarity measure of hesitant fuzzy sets are given.

Definition 3 (4). —

Sets A, B be the two hesitant fuzzy sets defined on X={x1, x2,…, xm}, and then the distance measure between A and B satisfies the following conditions:

  1. 0 ≤ d(A, B) ≤ 1

  2. d(A, B)=0, if and only if hA(x)=hB(x),  ∀ xX

  3. d(A, B)=d(B, A)

Definition 4 (4). —

Sets A, B be the two hesitant fuzzy sets defined on X={x1, x2,…, xm}, and then the similarity measure between A and B satisfies the following conditions:

  1. 0 ≤ S(A, B) ≤ 1

  2. S(A, B)=1, if and only if hA(x)=hB(x),  ∀ xX

  3. S(A, B)=S(B, A)

Definition 2.3 is proposed to facilitate the use of distance measures to define similarity measures. In practice, the distance can only be satisfied with d(A, B) ≥ 0.

3. A New Kind of Hesitant Fuzzy Entropy

Entropy is the measurement of the degree of uncertainty of information, and it has always been an important research object in uncertainty decision analysis. A new hesitant fuzzy entropy measure function is proposed by analyzing the shortcomings in the current research results on hesitant fuzzy entropy.

Definition 5 . —

Assign the hesitant fuzzy element h=∪γh{γ}={γj}j=1l, where l is the number of element in the hesitant fuzzy element, and record

x=1lj=1l12γj0.5,y=2ll1i,j=1,i<jlγiγj,l>1,0,l=1, (3)

where x represents the fuzzy degree of the hesitant fuzzy element h and y represents the hesitant degree of the hesitant fuzzy element h. Then the real valued function E : H⟶[0,1] on the hesitant fuzzy element h can be expressed by a binary function E(x, y), if the E(x, y) meets the following conditions:

  1. E(x, y)=0 if and only if x=0 and y=0

  2. E(x, y)=1 if and only if (x, y)=(1,0) or (x, y)=(0,1)

  3. E/∂x > 0 and ∂E/∂y > 0, ∂2E/∂x2 > 0 and ∂2E/∂y2 > 0

  4. E(x, y)=E(y, x)

Then, E(x, y) can be called as a hesitant fuzzy entropy function.

3.1. Interpretation and Analysis

  1. y=0⟷l=1, x=0⟷h={0} ∨ h={1} ∨ h={0,1}. Then x=0 ∧ y=0⟷h={0} ∨ h={1} indicates that h is a clear set, then the entropy is 0.

  2. When l > 1, since x=(1/l)∑j=1l(1 − 2|γj − 0.5|)=(1/l(l − 1))∑i,j=1,i<jl[(1 − 2|γi − 0.5|)+(1 − 2|γj − 0.5|)], y=(2/l(l − 1))∑i,j=1,i<jl|γiγj|=(2/l(l − 1))∑i,j=1,i<jl|(γi − 0.5)+(0.5 − γj)| ≤ (2/l(l − 1))∑i,j=1,i<jl|γi − 0.5)|+|0.5 − γj|, then it can get x+y ≤ 1, and it is concluded that the domain of the entropy function E(x, y) is {(x, y) | x ≥ 0, y ≥ 0, x+y ≤ 1} because the entropy function E(x, y) is concave increase with respect to x and y, and the maximum value of E(x, y) is 1 when (x, y)=(1,0) or (x, y)=(0,1) is obtained; that is, when h={0.5} or h={0,1}, the uncertainty reaches the maximum. h={0,1} is completely contradictory information, and h={0.5} is completely fuzzy information; in both cases, the uncertainty is maximized and in line with intuitive judgment.

  3. It ensures that the entropy function is concavely increased with respect to fuzziness and hesitation degree, conforms to human cognitive characteristics, and improves discrimination.

  4. Fuzziness and hesitancy have the same effect on entropy.

Based on the above analysis, function E(x, y)=x2+y2 obviously satisfies the above conditions in Definition 3.1, so it can be regarded as an entropy function. For example, if h={0.5}⟶x=1, y=0, then E(h)=E(x, y)=E(1,0)=1; if h={0,1}⟶x=0, y=1, then E(h)=E(x, y)=E(0,1)=1; and if h1={0.2}⟶x=0.4, y=0, h1={0.1, 0.2, 0.3}⟶x=0.4, y=0.13, then E(h1)=E(0.4, 0)=0.16, E(h2)=E(0.4, 0.13)=0.16+0.017=0.177, where E(h1) < E(h2). The above judgment results are consistent with the intuition.

Property 1 . —

Set hesitant fuzzy element h=∪γh{γ}={γj}j=1l, when l=1; the hesitant fuzzy element h degenerates into a fuzzy number, and the entropy of fuzzy value h is E(h)=E(x, y).

Proof —

  1. E(h)=E(x, y)=0⟺x=0, y=0, that is, h={0} or h={1}, where h is a clear set.

  2. According to condition (2) E(h)=E(x, y)=1⟺(x, y)=(0,1) or (x, y)=(1,0) because when l=1, y=0, so E(h)=E(x, y)=1⟺(x, y)=(1,0), that is, h={1/2}.

  3. According to condition (3), it is known that E(x, y) increases monotonously with respect to x, so when γj is closer to 0.5, the larger the x=(1/l)∑j=1l(1 − 2|γj − 0.5|) is, the larger the entropy E(h)=E(x, y) of the fuzzy value h is.

The property 3.1 indicates that the fuzzy entropy is a special case of the hesitant fuzzy entropy function, and the hesitant fuzzy entropy function can also be applied to the fuzzy set.

In order to illustrate the advantage of the entropy function proposed in this paper in measuring uncertainty, the following is compared with the existing entropy formula: at present, the common formulas of hesitating fuzzy entropy include the entropy formula proposed by Xu and Xia and the entropy formula proposed by Farhadinia, in which

EXuh=1lln2i=1lγi+γli+12lnγi+γli+12+2γiγli+12ln2γiλli+12, (4)

where l indicates the number of elements in a hesitant fuzzy number h and γi indicates the element of the largest ith in the hesitant fuzzy number h.

EPah=Z2dh,0.5Z1Z0Z1, (5)

where Z : [0,1]⟶[0,1] is strictly monotonically decreasing function, which may get Z(t)=1 − t, Z(t)=(1 − t)/(1+t), Z(t)=1 − t2, Z(t)=1 − tet−1; d(h, {0.5})=(1/l)∑i=1l|γi − 0.5| (γih, where l indicates the number of elements contained in the fuzzy number h).

Set hesitant fuzzy number h1=H{0.2, 0.4}, h2=H{0.3, 0.5}, h3=H{0.1, 0.2, 0.3}, h4=H{0.4, 0.5, 0.6}, h5=H{0.3, 0.6}, h6=H{0.4, 0.5}, h7=H{0.3, 0.5, 0.6}, and h8=H{0.2, 0.5, 0.7}. The entropy formula proposed by Xu and Xia and the entropy formula proposed by Farhadinia are compared with the entropy function proposed in this paper. The results are shown in Table 1.

Because the entropy formula proposed by Farhadinia only considers fuzziness and neglects the influence of hesitancy, the result is quite different from that of the method proposed in this paper and the method proposed by Xu. It is not difficult to find from the above table that the method proposed in this paper is obviously higher in the discrimination than that proposed by Xu, and the comparison result is close to it, and the individual results are inconsistent. For example, EXu(h4) > EXu(h6); however, according to the method presented in this paper, the result is ELv(h4) < ELv(h6); this is because the starting point is inconsistent and the hesitant fuzzy entropy proposed by Xu requires that the number of elements contained in the two pairs be equal and that the elements should be artificially added when the number of elements is different. Therefore, the proposed method is bound to deviate from intuitive judgment for the comparison of entropy of two hesitant fuzzy numbers with a different number of elements contained. The entropy measure function proposed in this paper not only considers the influence of fuzziness on the entropy value but also considers the effect of hesitation degree on the entropy value, which can more reasonably depict the uncertainty degree of the hesitation fuzzy number, so the result is more consistent with our intuition.

Table 1.

Comparison table of hesitant fuzzy entropy values.

E(h) h 1 h 2 h 3 h 4 h 5 h 6 h 7 h 8
E Xu(h) 0.881 0.971 0.640 1 0.993 0.993 0.995 0.995
E Pa(h) 0.6 0.8 0.4 0.867 0.7 0.9 0.8 0.6
E Lv(h) 0.4 0.68 0.177 0.774 0.58 0.82 0.68 0.556

4. Hesitant Fuzzy Distance Measure and Similarity Measure

For a hesitant fuzzy element h=∪γh{γ}={γj}j=1l, the most important information it contains is the size of γ value and the degree of uncertainty, which is also a common concern in practical applications. Based on this, we introduce the definition of feature vector of hesitant fuzzy information.

Definition 6 . —

Set hesitant fuzzy element h=∪γh{γ}={γj}j=1l; two-dimensional vector (s(h), E(h)) is called the information feature vector of hesitant fuzzy element h, which is marked as

h=sh,Eh, (6)

where s(h)=(1/l)∑j=1lγj represents the size of a hesitant fuzzy element and E(h) is the entropy of hesitant fuzzy element h, representing its degree of uncertainty, calculated by Definition 3.1.

The number of elements contained in different hesitant fuzzy elements may be different. In order to facilitate sorting and measurement, it is usually necessary to add elements artificially, which will inevitably destroy the original structure of the data and change the data information. The hesitant fuzzy element is proposed by the information feature vector to solve this kind of problem. The following formulas of the measure and similarity measure of hesitation fuzzy distance based on the feature vector of hesitant fuzzy information are given.

The feature vector of hesitant fuzzy information is to describe the information feature of the hesitant fuzzy element from two different factors, so the dimensions of different components are different; at the same time, there is obviously a correlation between the two components. Therefore, it is not appropriate to choose the traditional distance formula to measure the difference between the two hesitant fuzzy elements. This paper defines a new measure of distance and similarity from the angle of information theory.

Definition 7 . —

Set X={x1, x2,…, xm} as a nonempty domain; Aj={<xi, hAj(xi) > ∣xiX}, j=1,2, is the two hesitant fuzzy sets defined on X, and its information feature vectors are separately Aj={(sAj(xi), EAj(xi))∣xiX}, j=1,2. For the convenience of writing, note (sAj(xi), EAj(xi))=(sij, Eij)=Aj(xi), then call

dA1,A2=RA1,A2+RA2,A1, (7)

as the distance measure of A1, A2. Among which,

RA1,A2=1mi=1mΔ=s,EΔi1·logΔi1Δi2+1Δi1·log1Δi11Δi2,RA2,A1=1mi=1mΔ=s,EΔi2·logΔi2Δi1+1Δi2·log1Δi21Δi1, (8)

where Δ is the symbolic variable, Δ ∈ {s, E}.

The distance measure based on the information feature vector is based on the relative entropy idea, and it is easy to verify that it satisfies the following properties.

Property 2 . —

Set X={x1, x2,…, xm} as a nonempty domain; Aj={<xi, hAj(xi) > ∣xiX}, j=1,2, is the two hesitant fuzzy sets defined on X, and its distance measure is d(A1, A2), then

  1. d(A1, A2) ≥ 0

  2. d(A1, A2)=0, if and only if A1(xi)=A2(xi), i=1,2,…, m

  3. d(A1, A2)=d(A2, A1)

Proof —

Makef(t)=−log t; according tof(t)=(1/t2) > 0, f(t) is the concave function in a defined domain, that is, f(λ1t1+λ2t2) ≤ λ1f(t1)+λ2f(t2), among which, λ1, λ2 ∈ (0,1) and λ1+λ2=1. If and only if t1=t2, the equal sign is established. Suppose t1=(Δi2i1), t2=(1 − Δi2/1 − Δi1), λ1i1, λ2=1 − Δi1, bring in f(λ1t1+λ2t2) ≤ λ1f(t1) +λ2f(t2) and get

0Δi1·logΔi1Δi2+1Δi1·log1Δi11Δi2. (9)

If and only if when t1=(Δi2i1)=(1 − Δi2/1 − Δi1)=t2, Δi1 · log(Δi1i2)+(1 − Δi1) · log(1 − Δi1/1 − Δi2)=0; at this moment, Δi1i2. Then, R(A1, A2)=(1/m)∑i=1m(∑Δ=s,Ei1 · log(Δi1i2)+(1 − Δi1) · log(1 − Δi1/1 − Δi2)])≥0; if and only if when Δi1i2(Δ=s, E, i=1,2,…, m), R(A1, A2)=0.

In the same way, we can get R(A2, A1) ≥ 0; if and only if when Δi1i2(Δ=s, E, i=1,2,…, m), R(A2, A1)=0. In summary, (1) and (2) can be established.

According to the expression itself, it can be judged that (3) is clearly established.

Note. —

Property (2), A1(xi)=A2(xi), i=1,2,…, m, is not equivalent to A1=A2; for example, take A1={<x, {0.5}>} and A2={<x, {0,1}>}, it is obvious that A1A2. The information feature vector is represented as A1={<x, (0.5, 1)>} and A2={<x, {0.5, 1}>}, respectively, according to property (2), and then d(A1, A2)=0. At this moment, the result is consistent with human intuition because completely ambiguous information and completely contradictory information can convey the same amount of information. This is also the main difference between the distance measure proposed in this paper and other hesitant fuzzy distance measures.

Inspired by TOPSIS, the hesitant fuzzy similarity measure formula based on hesitant fuzzy distance measure is given below.

Definition 8 . —

Let X={x1, x2,…, xm} as a given nonempty domain; j=1,2 is the two hesitant fuzzy sets defined on X, and j=1,2 modified to A1 and A2.

SA1,A2=dA1,A2cdA1,A2+dA1,A2c, (10)

as the similarity measures of A1, A2.

Property 3 . —

Set X={x1, x2,…, xm} as a nonempty domain; Aj={<xi, hAj(xi) > ∣xiX}, j=1,2, is the two hesitant fuzzy sets defined on X, and the similarity measure A1, A2 of is S(A1, A2):

  1. 0 ≤ S(A1, A2) ≤ 1

  2. S(A1, A2)=S(A2, A1)

  3. S(A1, A2)=0, if and only if when A1(xi)=A2c(xi), i=1,2,…, m

  4. S(A1, A2)=1, if and only if when A1(xi)=A2(xi), i=1,2,…, m

  5. If and only if when d(A1, A2)=d(A1, A2c), S(A1, A2)=(1/2)

Property 4.2 can be determined by the formula itself. The proof process is omitted.

In practical application, different elements in set X have different status and should be given different weights. A similarity measure formula considering weights is given below:

SwA1,A2=dwA1,A2cdwA1,A2+dwA1,A2c, (11)

among which dw(A1, A2)=Rw(A1, A2)+Rw(A2, A1)=∑i=1m(wiΔ=s,Ei1 · log(Δi1i2)+(1 − Δi1) · log((1 − Δi1)/(1 − Δi2))])+∑i=1m(wiΔ=s,Ei2 · log(Δi2i1)+(1 − Δi2) · log((1 − Δi2)/(1 − Δi1))]). wi is the weight of element xi(i=1,2,…, m) and satisfies ∑i=1mwi=1, wi ∈ [0,1]. Obviously, when wi=(1/m)(i=1,2,…, m), dw(A1, A2)=d(A1, A2) and Sw(A1, A2)=S(A1, A2).

5. Network Clustering Method Based on Hesitant Fuzzy Similarity Measure

The network clustering [33] method is a common method in data clustering analysis, and it is also the best choice to extend the clustering method to the fuzzy environment. The specific process is as follows: the similarity coefficient matrix P is constructed by the data similarity measure, and then the cutting level λ ∈ [0,1] is selected as λ−truncated matrix Pλ of P, and replace the principal diagonal element with the scheme symbol. In the lower left of the principal diagonal, the symbol “∗” is used instead of “1” to remove the “0” element. The position of the symbol “∗” is called the node. The so called network is to cross the nodes as the latitude and longitude lines and tie the scheme corresponding to the latitude and longitude lines at the nodes to achieve classification. The main advantage is that the clustering results can be obtained quickly and effectively by using the similarity coefficient matrix directly on the table. The method of clustering analysis in the hesitant fuzzy environment is given below. The calculation process is as follows:

  1. Let A={A1, A2,…, Am} be the set of object to be classified, F={F1, F2,…, Fn} be the decision factor set, and W=(w1, w2,…,wn)T be the decision factor weight vector. The decision expert group measures the classified objects according to the decision factors and obtains the hesitant fuzzy decision matrix D=(hij)m×n.

  2. According to formula (6), the hesitant fuzzy value hij is expressed by the information feature vector, and then the decision matrix D is transformed into the information feature vector matrix.

  3. Calculate the hesitant fuzzy similarity coefficient matrix P=(Sw(Ai, Ak))m×m by using formula (11).

  4. Remove elements above the principal diagonal and replace the principal diagonal element with the scheme symbol.

  5. Select cutting level λ ∈ [0,1] as the λ−truncated matrix Pλ of P, in the lower left of the principal diagonal, the symbol “∗” is replaced by “1,” and the “0” element is removed. The position of the symbol “∗” is called a node, the node is the latitude and longitude lines, and the node is over the node. The schemes corresponding to the latitude and longitude lines are bundled into one category.

6. Illustrative Example

In order to facilitate comparative analysis, this paper uses an example from the literature [32]. Through four factors (price F1, function F2, after-sales service F3, and quality F4), 7 cell phones Ai(i=1,2,…, 7) are classified. Assume the factor weight vector is w=(0.3, 0.25, 0.2, 0.25)T. The decision group gives the evaluation value of mobile phone Ai under the decision factor Fj, which is represented by the hesitant fuzzy set Ai={hij | j=1,2,3,4}, (i=1,2,…, 7), among which hij indicates the degree to which the mobile Ai satisfies the decision factor Fj. Then the decision information can be represented by the decision matrix D=(hij)7×4 (Table 2). According to the network clustering method, cluster analysis is performed on 7 mobile phones as shown in Table 2.

Table 2.

Hesitation fuzzy decision matrix.

F 1 F 2 F 3 F 4
A 1 {0.4, 0.6} {0.3, 0.5, 0.6} {0.2, 0.3} {0.4}
A 2 {0.2, 0.3, 0.5} {0.2, 0.4} {0.2, 0.4, 0.5} {0.3, 0.4}
A 3 {0.2, 0.3, 0.4} {0.2, 0.4} {0.2, 0.3, 0.5} {0.3}
A 4 {0.3, 0.45, 0.75} {0.3, 0.6} {0.3, 0.6, 0.75} {0.45, 0.6}
A 5 {0.8} {0.4, 0.6, 0.7} {0.3, 0.4} {0.6, 0.8}
A 6 {0.4, 0.5, 0.7} {0.4, 0.6} {0.4, 0.6, 0.7} {0.5, 0.6}
A 7 {0.6, 0.7, 0.9} {0.6, 0.8} {0.6, 0.8, 0.9} {0.7, 0.8}

Step 1 . —

See Table 2.

Step 2 . —

According to the formula (6), the hesitation fuzzy value in the hesitating fuzzy decision matrix is transformed into the information feature vector matrix (Table 3).

For example, the data (0.5, 0.68) in the first column of the first row in Table 3 are the information feature vector corresponding to the data {0.4, 0.6} in the first column of the first row in Table 2. They are calculated according to formula (6), where 0.5=(1/2)(0.4+0.6), 0.68={(1/2) · [(1 − 2  · |0.4 − 0.5|)+(1 − 2 · |0.6 − 0.5|)]}2+(|0.4 − 0.6|)2.

Table 3.

Information feature vector matrix.

F 1 F 2 F 3 F 4
A 1 (0.5, 0.68) (0.47, 0.68) (0.25, 0.26) (0.4, 0.64)
A 2 (0.33, 0.48) (0.3, 0.4) (0.37, 0.57) (0.35, 0.5)
A 3 (0.3, 0.38) (0.3, 0.4) (0.33, 0.48) (0.3, 0.36)
A 4 (0.5, 0.53) (0.45, 0.58) (0.55, 0.49) (0.53, 0.75)
A 5 (0.8, 0.16) (0.57, 0.57) (0.35, 0.5) (0.7, 0.4)
A 6 (0.53, 0.68) (0.5, 0.68) (0.57, 0.58) (0.55, 0.82)
A 7 (0.73, 0.32) (0.7, 0.4) (0.77, 0.26) (0.75, 0.26)

Step 3 . —

Calculate hesitant fuzzy similarity coefficient matrix by using formula (4):

P=10.61830.61680.42980.51210.40770.36960.618310.92920.50200.30640.44340.19850.61680.929210.49270.24840.45490.07930.42980.50200.492710.49340.56530.51390.51210.30640.24840.493410.51690.71720.40770.44340.45490.56530.516910.54010.36960.19850.07930.51390.71720.54011. (12)

The first row and second column data 0.6183 are the similarity measurement between the date in the first row and the date in the second row in Table 3 and is calculated according to formula (11):

dwA1,A2=0.3·0.5·log0.50.33+10.5·log10.510.33+0.68·log0.680.48+10.68·log10.6810.48+0.33·log0.330.5+10.33·log10.3310.5+0.48·log0.480.68+10.48·log10.4810.68+0.25·0.47·log0.470.3+10.47·log10.4710.3+0.68·log0.680.4+10.68·log10.6810.4+0.3·log0.30.47+10.3·log10.310.47+0.68·log0.680.4+10.68·log10.6810.4+0.2·0.25·log0.250.37+10.25·log10.2510.37+0.26·log0.260.57+10.26·log10.2610.57+0.37·log0.370.25+10.37·log10.3710.25+0.57·log0.570.26+10.57·log10.5710.26+0.25·0.4·log0.40.35+10.4·log10.410.35+0.64·log0.640.5+10.64·log10.6410.5+0.35·log0.350.4+10.35·log10.3510.4+0.5·log0.50.64+10.5·log10.510.64=0.3169,dwA1,A2c=0.3·0.5·log0.50.67+10.5·log10.510.67+0.68·log0.680.48+10.68·log10.6810.48+0.67·log0.670.5+10.67·log10.6710.5+0.48·log0.480.68+10.48·log10.4810.68+0.25·0.47·log0.470.7+10.47·log10.4710.7+0.68·log0.680.4+10.68·log10.6810.4+0.7·log0.70.47+10.7·log10.710.47+0.68·log0.680.4+10.68·log10.6810.4+0.2·0.25·log0.250.63+10.25·log10.2510.63+0.26·log0.260.57+10.26·log10.2610.57+0.63·log0.630.25+10.63·log10.6310.25+0.57·log0.570.26+10.57·log10.5710.26+0.25·0.4·log0.40.65+10.4·log10.410.65+0.64·log0.640.5+10.64·log10.6410.5+0.65·log0.650.4+10.65·log10.6510.4+0.5·log0.50.64+10.5·log10.510.64=0.5134. (13)

Step 4 . —

Remove elements above the principal diagonal, and replace the principal diagonal element with the scheme symbol, that is:

P=A10.6183A20.61680.9292A30.42980.50200.4927A40.51210.30640.24840.4934A50.40770.44340.45490.56530.5169A60.36960.19850.07930.51390.71720.5401A7. (14)

Step 5 . —

Select the cutting level λ ∈ [0,1] as theλ−truncation matrix Pλ of P and then classify through the network:

  1. When 0.9292 < λ ≤ 1, they are divided into 7 categories: {A1}, {A2}, {A3}, {A4}, {A5}, {A6}, {A7}

  2. When 0.7172 < λ ≤ 0.9291, they are divided into 6 categories: {A1}, {A2, A3}, {A4}, {A5}, {A6}, {A7}

  3. When 0.6183 < λ ≤ 0.7172, they are divided into 5 categories: {A1}, {A2, A3}, {A4}, {A5, A7}, {A6}

  4. When 0.5653 < λ ≤ 0.6183, they are divided into 4 categories: {A1, A2, A3}, {A4}, {A5, A7}, {A6}

  5. When 0.5401 < λ ≤ 0.5653, they are divided into 3 categories: {A1, A2, A3}, {A4, A6}, {A5, A7}

  6. When 0.4927 < λ ≤ 0.5401, they are divided into 2 categories: {A1, A2, A3}, {A4, A5, A6, A7}

  7. When 0 ≤ λ ≤ 0.4927, they are divided into 1 category: {A1, A2, A3, A4, A5, A6, A7}

Next, the clustering results of this paper are compared with those of the literature [28], literature [29], literature [31], and literature [32], and the results are analyzed. The clustering results obtained by other methods are shown in Table 4.

Different literatures choose different measures to measure the degree of closeness between samples; among them, the literature [28] is based on the correlation coefficient; the literature [29] and literature [31] are based on the distance measure; and the literature [32] and this paper are based on the similarity measure. In order to compare the sensitivity of various measures, it is necessary to analyze the variance of the measurement data. The larger the variance, the higher the sensitivity of the corresponding measure. The results are shown in Table 5. Furthermore, in order to compare the effectiveness of various methods, the D-B index [34] of the clustering results is calculated separately. The distance measure selected for calculating the D-B index of the literature [28] is d(A, B)=1 − ρ(A, B). The results are shown in Table 5.

It can be found from Table 5 that (1) the similarity measure proposed in this paper has higher sensitivity than other measures and the clustering result has better robustness and (2) tThe D-B index of this paper is smaller, indicating that the clustering results are better.

Through comparison, it can be found that (1) the results obtained by using the method proposed in this paper are consistent with those obtained in the literature [32], which to some extent reflects the effectiveness of the method proposed in this paper; (2) the results of the classification in the literature [28] are not precise and accurate, as it is intuitively possible to judge from the data in Table 2S(A2, A3) > S(A2, A4), so it is more appropriate to classify them A2, A3 as a group; (3) the methods of the literature [29] and literature [31] are put forward based on the distance formula in the literature [4], but the resolution of the distance formula is not high, and the results are sometimes inconsistent with the facts [32], which will inevitably affect the classification results; (4) the data must be processed by equal length in the literature [28], literature [29], literature [31], and literature [32], which will inevitably affect the clustering results. The reason why the results in this paper are consistent with the literature [32] is that the difference in the number of hesitant fuzzy numbers in Table 2 is small, and the size of each element in the same hesitant fuzzy number is not much different, and if not so, the results must be different from those obtained by the present method.

Table 4.

Clustering result comparison table.

Category Literature [28] method Literature [29] method Literature [31] method Literature [32] method
1 {A1, A2, A3, A4, A5, A6, A7} {A1, A2, A3, A4, A5, A6, A7} {A1, A2, A3, A4, A5, A6, A7} {A1, A2, A3, A4, A5, A6, A7}
2 {A1, A2, A3, A4, A5, A6}, {A7} {A1, A2, A3, A4, A6, A7}, {A5} {A1, A2, A3}, {A4, A5, A6, A7}
3 {A1, A5}, {A2, A3, A4}, {A6, A7} {A1, A2, A3, A4, A6}, {A5}, {A7} {A1, A2, A3, A4, A6}, {A5}, {A7} {A1, A2, A3}, {A4, A6}, {A5, A7}
4 {A1}, {A2, A3, A4}, {A5}, {A6, A7} {A1, A4, A6}, {A2, A3}, {A5}, {A7} {A1, A2, A3}, {A4, A6}, {A5}, {A7} {A1, A2, A3}, {A4}{A5, A7}, {A6}
5 {A1}, {A2, A4}, {A3}, {A5}, {A6, A7} {A1}, {A2, A3}, {A4, A6}, {A5}, {A7} {A1}, {A2, A3}, {A4, A6}, {A5}, {A7} {A1}, {A2, A3}, {A4}, {A5, A7}, {A6}
6 {A1}, {A2, A4}, {A3}, {A5}, {A6}, {A7} {A1}, {A2, A3}, {A4}, {A5}, {A6, A7} {A1}, {A2, A3}, {A4}, {A5}, {A6, A7} {A1}, {A2, A3}, {A4}, {A5}, {A6}, {A7}
7 {A1}, {A2}, {A3}, {A4}, {A5}, {A6}, {A7} {A1}, {A2}, {A3}, {A4}, {A5}, {A6}, {A7} {A1}, {A2}, {A3}, {A4}, {A5}, {A6}, {A7}

Table 5.

Variance comparison table.

Literature [28] Literature [29] Literature [31] Literature [32] This paper
Variance 0.0038 0.0152 0.0152 0.0319 0.0328
D-B index 1.7652 0.9236 0.9983 0.7326 0.6537

7. Conclusion

In this paper, the hesitant fuzzy information feature vector is used as the entry point, which provides a new idea for solving various hesitant fuzzy measures. Then the hesitant fuzzy uncertainty measure, distance measure, and similarity measure are studied. Finally, a clustering method for fuzzy information is proposed. Through analyzing the results of the example, it has been proved that this method is faster and more effective in practical applications. The main contributions of this paper are (1) it effectively avoids the problem of processing data with equal length in the research of the measure of hesitant fuzzy set and (2) combined with the similarity measure proposed by TOPSIS idea, the resolution between schemes can be improved. Subsequent research on hesitant fuzzy set theory and application based on hesitant fuzzy information feature vectors will be a meaningful topic.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (no. 61350003).

Data Availability

Previously reported data were used to support this study and are available at DOI: 10.1080/00207721.2013.797073. These prior studies (and datasets) are cited at relevant places within the text as references [29].

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  • 1.Torra V., Narukawa Y. On hesitant fuzzy sets and decision. Proceedings of 18th IEEE International Conference on Fuzzy Systems; August 2009; Jeju Island, Korea. pp. 1378–1382. [Google Scholar]
  • 2.Torra V. Hesitant fuzzy sets. International Journal of Intelligent Systems. 2010;25:529–539. [Google Scholar]
  • 3.Zadeh L. A. Fuzzy sets. Information and Control. 1965;8(3):338–353. doi: 10.1016/s0019-9958(65)90241-x. [DOI] [Google Scholar]
  • 4.Xu Z., Xia M. Distance and similarity measures for hesitant fuzzy sets. Information Sciences. 2011;181(11):2128–2138. doi: 10.1016/j.ins.2011.01.028. [DOI] [Google Scholar]
  • 5.Xu Z., Xia M. On distance and correlation measures of hesitant fuzzy information. International Journal of Intelligent Systems. 2011;26(5):410–425. doi: 10.1002/int.20474. [DOI] [Google Scholar]
  • 6.Farhadinia B. Information measures for hesitant fuzzy sets and interval-valued hesitant fuzzy sets. Information Sciences. 2013;240:129–144. doi: 10.1016/j.ins.2013.03.034. [DOI] [Google Scholar]
  • 7.Li D., Zeng W., Li J. New distance and similarity measures on hesitant fuzzy sets and their applications in multiple criteria decision making. Engineering Applications of Artificial Intelligence. 2015;40:11–16. doi: 10.1016/j.engappai.2014.12.012. [DOI] [Google Scholar]
  • 8.Li D., Zeng W., Zhao Y. Note on distance measure of hesitant fuzzy sets. Information Sciences. 2015;321:103–115. doi: 10.1016/j.ins.2015.03.076. [DOI] [Google Scholar]
  • 9.Hesamian G., Shams M. Measuring similarity and ordering based on hesitant fuzzy linguistic term sets. Journal of Intelligent and Fuzzy Systems. 2015;28(2):983–990. [Google Scholar]
  • 10.Li C. C., Liu L. Z. New distance measures on hesitant fuzzy sets. Fuzzy Systems and Mathematics. 2018;32:114–124. [Google Scholar]
  • 11.Xia M. M., Xu Z. S. Hesitant fuzzy information aggregation in decision making. International Journal of Approximate Reasoning. 2015;52:395–407. [Google Scholar]
  • 12.Zhang Z. Hesitant fuzzy power aggregation operators and their application to multiple attribute group decision making. Information Sciences. 2013;234:150–181. doi: 10.1016/j.ins.2013.01.002. [DOI] [Google Scholar]
  • 13.Meng F., Chen X., Zhang Q. Induced generalized hesitant fuzzy Shapely hybrid operators and their application in multi-attribute decision making. Applied Soft Computing. 2015;28:599–607. doi: 10.1016/j.asoc.2014.11.017. [DOI] [Google Scholar]
  • 14.Torres R., Salas R., Astudillo H. Time-based hesitant fuzzy information aggregation approach for decision-making problems. International Journal of Intelligent Systems. 2014;29(6):579–595. doi: 10.1002/int.21658. [DOI] [Google Scholar]
  • 15.Tan C., Yi W., Chen X. Hesitant fuzzy Hamacher aggregation operators for multicriteria decision making. Applied Soft Computing. 2015;26:325–349. doi: 10.1016/j.asoc.2014.10.007. [DOI] [Google Scholar]
  • 16.Wei C., Zhao N., Tang X. Operators and comparisons of hesitant fuzzy linguistic term sets. IEEE Transactions on Fuzzy Systems. 2014;22(3):575–585. doi: 10.1109/tfuzz.2013.2269144. [DOI] [Google Scholar]
  • 17.Chen N., Xu Z., Xia M. Interval-valued hesitant preference relations and their applications to group decision making. Knowledge-Based Systems. 2013;37:528–540. doi: 10.1016/j.knosys.2012.09.009. [DOI] [Google Scholar]
  • 18.Zhu B., Xu Z., Xia M. Dual hesitant fuzzy sets. Journal of Applied Mathematics. 2012;2012:13. doi: 10.1155/2012/879629.879629 [DOI] [Google Scholar]
  • 19.Rodriguez R. M., Martinez L., Herrera F. Hesitant fuzzy linguistic term sets for decision making. IEEE Transactions on Fuzzy Systems. 2012;20(1):109–119. doi: 10.1109/tfuzz.2011.2170076. [DOI] [Google Scholar]
  • 20.Wang H. Extended hesitant fuzzy linguistic term sets and their aggregation in group decision making. International Journal of Computational Intelligence Systems. 2015;8(1):14–33. doi: 10.1080/18756891.2014.964010. [DOI] [Google Scholar]
  • 21.Yang X. B., Song X. N., Qi Y. S. Constructive and axiomatic approaches to hesitant fuzzy rough set. Soft Computing. 2014;18(6):1067–1077. doi: 10.1007/s00500-013-1127-2. [DOI] [Google Scholar]
  • 22.Wang F. Q., Li X. H., Chen X. H. Hesitant fuzzy soft set and its applications in multicriteria decision making. Journal of Applied Mathematics. 2014;2014:10. doi: 10.1155/2014/643785.643785 [DOI] [Google Scholar]
  • 23.Garmendia L., González del Campo R., Recasens J. Partial orderings for hesitant fuzzy sets. International Journal of Approximate Reasoning. 2017;84:159–167. doi: 10.1016/j.ijar.2017.02.008. [DOI] [Google Scholar]
  • 24.Aliahmadipour L., Torra V., Eslami E. Fuzzy Sets, Rough Sets, Multisets and Clustering. Vol. 671. Berlin, Germany: Springer; 2017. On hesitant fuzzy clustering and clustering of hesitant fuzzy data; pp. 157–167. [Google Scholar]
  • 25.Wu X. H., Wu B., Sun J., Zhao J. W. Mixed fuzzy inter-cluster separation clustering algorithm. Applied Mathematical Modelling. 2011;35(10):4790–4795. doi: 10.1016/j.apm.2011.03.050. [DOI] [Google Scholar]
  • 26.Wang Z., Xu Z. S., Liu S. S., Tang J. A netting clustering analysis method under intuitionistic fuzzy environment. Applied Soft Computing. 2011;11(8):5558–5564. doi: 10.1016/j.asoc.2011.05.004. [DOI] [Google Scholar]
  • 27.Hwang C., Rhee F. C. H. Uncertain fuzzy clustering: interval type-2 fuzzy approach to C-Means. IEEE Transactions on Fuzzy Systems. 2007;15(1):107–120. doi: 10.1109/tfuzz.2006.889763. [DOI] [Google Scholar]
  • 28.Chen N., Xu Z. S., Xia M. M. Correlation coefficients of hesitant fuzzy sets and their application to clustering analysis. Applied Mathematical Modelling. 2013;37(4):2197–2211. doi: 10.1016/j.apm.2012.04.031. [DOI] [Google Scholar]
  • 29.Zhang X. L., Xu Z. S. Hesitant fuzzy agglomerative hierarchical clustering algorithms. International Journal of Systems Science. 2013;46(3):562–576. doi: 10.1080/00207721.2013.797073. [DOI] [Google Scholar]
  • 30.Miyamoto S. Fuzzy Sets in Information Retrieval and Cluster Analysis. Dordrecht, The Netherlands: Kluwer; 1990. [Google Scholar]
  • 31.Zhang X. L., Xu Z. S. A MST clustering analysis method under hesitant fuzzy environment. Control and Cybernetics. 2012;41(3):645–666. [Google Scholar]
  • 32.Xiao-di L., Zhu J.-jun, Liu Si-feng. Similarity measure of hesitant fuzzy sets based on symmetric cross entropy and its application in clustering analysis. Control and Decision. 2014;29:1816–1822. [Google Scholar]
  • 33.He Z. X. Fuzzy Mathematics and Its Application. Tianjin, China: Tianjin Science and Technology Press; 1983. [Google Scholar]
  • 34.Kalyani M., Sushmita M. Clustering and its validation in a symbolic framework. Pattern Recognition. 2003;24(14):2367–2376. doi: 10.1016/s0167-8655(03)00066-7. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Previously reported data were used to support this study and are available at DOI: 10.1080/00207721.2013.797073. These prior studies (and datasets) are cited at relevant places within the text as references [29].


Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES