Abstract
Background
Biological data comprises various topologies or a mixture of forms, which makes its analysis extremely complicated. With this data increasing in a daily basis, the design and development of efficient and accurate statistical methods has become absolutely necessary. Specific analyses, such as those related to genome-wide association studies and multi-omics information, are often aimed at clustering sub-conditions of cancers and other diseases. Hierarchical clustering methods, which can be categorized into agglomerative and divisive, have been widely used in such situations. However, unlike agglomerative methods divisive clustering approaches have consistently proved to be computationally expensive.
Results
The proposed clustering algorithm (DRAGON) was verified on mutation and microarray data, and was gauged against standard clustering methods in the literature. Its validation included synthetic and significant biological data. When validated on mixed-lineage leukemia data, DRAGON achieved the highest clustering accuracy with data of four different dimensions. Consequently, DRAGON outperformed previous methods with 3-,4- and 5-dimensional acute leukemia data. When tested on mutation data, DRAGON achieved the best performance with 2-dimensional information.
Conclusions
This work proposes a computationally efficient divisive hierarchical clustering method, which can compete equally with agglomerative approaches. The proposed method turned out to correctly cluster data with distinct topologies. A MATLAB implementation can be extraced from http://www.riken.jp/en/research/labs/ims/med_sci_math/ or http://www.alok-ai-lab.com
Electronic supplementary material
The online version of this article (doi:10.1186/s12859-017-1965-5) contains supplementary material, which is available to authorized users.
Keywords: Divisive approach, Hierarchical clustering, Maximum likelihood
Background
In unsupervised clustering algorithms, the class label or the state of nature of a sample is unknown. The partitioning of data is then driven by considering similarity or distance measures. In some applications (e.g. genome-wide association studies, multi-omics data analyses), the number of clusters also remains unknown. Because such biological information usually tends to follow a normal distribution, the distribution of samples of each cluster can be assumed to be Gaussian.
Hierarchical clustering methods, which can be mainly categorized into agglomerative (bottom-up) and divisive (top-down) procedures, are well known [1–20]. In agglomerative procedures, each sample is initially assumed to be a cluster. The two nearest clusters (based on a distance measure or criterion function) are then merged at a time. This merger continues until all the samples are clustered into one group. Consequently, a tree like structure known as dendrogram is yielded. If the number of clusters is provided, the process of amalgamation of clusters can be terminated when the desired number of clusters is obtained. The first step of an agglomerative procedure considers all the possible mergers of two samples, which requires n(n − 1)/2 combinations (where n depicts the number of samples). Divisive procedures, on the other hand, perform clustering in an inverse way as compared to their agglomerative counterparts. They begin by considering a group (having all the samples) and divide it into two groups at each stage until all the groups comprise of only a single sample [21, 22]. In the first step of a divisive procedure all the partitions of a sample set are considered, which amounts to 2n − 1 − 1 combinations. This number of combinations grows exponentially and practically makes divisive clustering a difficult procedure to implement. However, there are a few divisive approaches which do not necessarily consider all the divisions [21]. In hierarchical classifications, each subcluster can be formed from one larger cluster split into two, or the union of two smaller clusters. In either case, false decisions made in early stages cannot be corrected later on. For this reason, divisive procedures, which start with the entire dataset, are in general considered safer than agglomerative approaches [21, 23]. Therefore, the accuracy of a divisive procedure is envisaged to be higher than that of an agglomerative procedure [24]. However, the high computational demand (O(2n)~O(n 5)) of divisive procedures has severely restricted their usage [24, 25] (though for special cases the complexity can be further reduced [26]). Therefore, the divisive procedure has not been generally used for hierarchical clustering, remaining largely ignored in the literature.
Hierarchical approaches do not require initial parameter settings and generally employed either linear or non-linear regression models [27–29]. Over the last few decades, a number of hierarchical approaches have been proposed. Some of these popular schemes are summarized below. The single linkage or link agglomerative hierarchical approach (SLink) [30] merges two adjacent neighbour groups. Euclidean distance for computing the proximity between two clusters. SLink is very sensitive to data location and occasionally generates groups in a long chain (called as chaining effect). This chaining effect can be reduced developing a method based on farthest distance. This was achieved by the complete linkage (CLink) hierarchical approach [2]. Nevertheless, CLink is also sensitive to outliers. Sensitiveness could be further decreased by the average linkage (ALink) hierarchical approach [31, 32]. ALink implements linking by using the average distance between two groups. In a similar way, the median linkage (MLink) hierarchical approach [33] regards median distance for linking. In Ward’s linkage (Wa-Link), clusters are merged based on the optimal value of an objective function [34]. In weighted average distance linkage (Wt-Link) hierarchical clustering [35, 36], the group sizes are not considered when computing average distances. Consequently, smaller groups will be assigned larger weights during the clustering process [35]. Similarly, model-based hierarchical clustering [4, 8] uses an objective function. Whereas the method in [8] follows a Bayesian analysis and uses both Dirichlet priors and multinomial likelihood function, the approach in [4] optimizes the distance between two GMMs. The number of group is previously defined. Most of these approaches are constructed using the agglomerative procedure, though their construction (with higher computational demand) is equally possible using the divisive procedure. Although divisive clustering is generally disregarded, some approaches like DIANA (DIvisive ANAlysis) program has been recently established [21]. In spite of well-established methods (i.e. EM algorithm [37, 38]) for estimating the parameters of a Gaussian mixture model, it is worth noting that hierarchical and expectation-maximization (EM) algorithms are very different in nature. The EM algorithm is an iterative optimization method, which requires prior knowledge of the number of clusters. It begins with a random choice of cluster centers and therefore returns different sets of clusters for distinct runs of the algorithm. Hierarchical clustering algorithms, on the other hand, do not require such prior knowledge and return a unique set of clusters. These advantages often make hierarchical clustering methods preferable to the EM algorithm for dealing with biological datasets where unique solutions are of utmost importance [39–41].
In this work, we described a new Divisive hieRArchical maximum likelihOod clusteriNg approach, abbreviated as DRAGON hereafter. This is a top-down procedure which does not find pairs. Instead, it takes out one sample at a time, maximally increasing the likelihood function. This process continues until the first cluster is obtained. This cluster is not further subdivided but removed from the sample set. In the remaining sample set, the same procedure is repeated for obtaining all the possible clusters. The removal of one sample out of n samples requires n search. This reduces the total search complexity to O(n 2 c) (where c is the number of clusters), which represents a significant reduction of the top-down procedure. The following sections present the mathematical derivation of the proposed model, and the analyses carried over on synthetic as well as on biological data to illustrate its usefulness.
Methods
Maximum likelihood clustering: an overview
This section summarizes an overview of the maximum likelihood method for clustering [22]. Here we are not introducing our method, instead we are proving a brief description of conventional maximum likelihood approach for clustering applications. It is possible to learn from an unlabeled data if some assumptions are taken. We will begin the section with an assumption that probability densities are known and it is required to estimate unknown parametric vector θ. The solution comes out to be similar to supervised learning case of maximum likelihood estimation. However, in the supervised learning case, the topology of groups of data is known. But in an unsupervised learning case one has to assume parametric form of data to reach to the solution. Here we describe how to estimate of maximum likelihood of clusters of a given sample set χ. The label of cluster of the sample sets is defined as ω. Assuming there are c clusters in the sample set (c ≥ 1), we define Ω = {ω j} (for j = 1, 2, …, c) as the cluster label for jth cluster χ j (In many clustering problems, the number of c is unknown, this issue we will deal in detail in later section and in Additional file 1). In this paper, we followed the notations from Duda et al. [22] for the convenience of readers. Let a sample set χ = {x 1, x 2, …, x n} be defined in a d-dimensional space (It is assumed that d < n. For d ≫ n, dimensionality reduction techniques can be first applied for supervised or unsupervised learning tasks [42–47]). Let an unknown parameter vector be θ consisting of mean μ as well as covariance Σ. This will specify the mixture density as
1 |
where p(x k|ω j, θ j) (for j = 1, …, c) is the conditional density, P(ω j) is the a priori probability and θ = {θ j}. The joint density is further defined using the log likelihood as
2 |
Assuming the joint density p(χ|θ) is differentiable w.r.t to θ then from Eqs. (1) to (2)
3 |
where is the gradient of L w.r.t. θ i. Assuming θ i and θ j are independent, and supposing a posteriori probability is
4 |
then from Eq. (4) we can observe that . Substituting this value in Eq. (3), we obtain
5 |
Note that in Eq. (5), (1/f(z))∇ z f(z) is arranged as ∇ z log f(z). Equation (5) can be equated to 0 () for obtaining maximum likelihood estimate . This will give us the solution as (interested readers may refer to Duda et al. [22] for further details.)
6 |
7 |
8 |
In the case of normal distribution, the unknown mean and covariance {μ, Σ} parameters are replaced in θ in the Eqs. 6, 7 and 8 for yielding maximum likelihood estimates. The parameter θ is usually updated in an iterative fashion to attain by EM algorithms or hill climbing schemes.
DRAGON method: concept
Here we illustrate the clustering method DRAGON. In brief, the proposed procedure is top-down in nature. It initially considers the sample set as one cluster from which one sample is removed at a time. This increases the likelihood function and continues until the maximum likelihood is reached as depicted in Fig. 1 (where L 1 is the cluster likelihood at the beginning of the process and L 3 is the maximum likelihood after removing two samples). Once the first cluster is obtained it is removed from the sample set and the procedure is then repeated for attaining the subsequent clusters. Consequently, only one cluster will be retrieved from a sample set. It is assumed that samples are multinomial distributed, however, the number of clusters is not known at the beginning of the process.
To establish the maximum likelihood estimate in the divisive hierarchical context, we investigate the criterion function and the distance measure that satisfy it.
DRAGON method: algorithm
To find the distance measure, we first define the log-likelihood function of a cluster χ s, where χ s is a subset of χ. At the beginning, χ s is the same as χ, however, in every subsequent iteration a sample is removed from χ s such that the likelihood function
9 |
is maximized.
Since we are finding only one cluster in the sample set χ, a priori probability P(ω) can be ignored. We would like to explore how function L changes when a sample is taken out. Let us suppose centroid μ and covariance Σ of χ s are defined as
10 |
11 |
where the number of samples in χ s is depicted as n. Assuming that the component density is normal then Eq. (9) can be simplified as
where trace function is denoted by tr(). Since , we can write L as
12 |
If a sample is removed from χ s then centroid and covariance (Eqs. (10) and (11)) will change as follows
13 |
14 |
In order to observe the alteration in the likelihood function (of Eq. (12)), we provide the following Lemma.
Lemma 1 Assume point is taken out of a set χ s and this changes the centroid and covariance (as Eqs. ( 13 ) and ( 14 ) described). Thereby the determinant of Σ * is defined as
Proof From Eq. (14), the determinant of Σ * will be
L1 |
For any square matrix of size m × m, we can write |AB| = |A||B|, and |cA| = c m|A| where c is any scalar, This would enable us to write Eq. (L1) in the following manner
L2 |
can be proved to be |I n × n + BA| by Sylvester’s determinant theorem (where A ∈ ℝm × n and B ∈ ℝn × m are rectangular matrices). This would allow us to write
For any scalar |c| = c, the Lemma is then proved by substituting this term in Eq. (L1). □
It is now possible to define the change in L as
15 |
where ΔL is defined as
16 |
and P is expressed as
17 |
It can be observed from Eqs. (15) to (17) that when a sample is taken out of cluster χ s, the change in L mainly depends on the term P as all the other terms are not changing. If we want to select such that L * > L, this requires to solve the following maximization problem
18 |
Therefore, by removing the likelihood should increase until the maximum value is reached.
This procedure can track the location of the cluster having the highest density or likelihood. Because one sample is taken out at a time, it could be sensitive to data positions around the center of the cluster. Thereby, it is possible to locate the center of the cluster whereas its complete topology can be missed. In order to reduce such sensitiveness additional processing for tuning the cluster would be useful.
By taking out one sample at a time, we can obtain a cluster χ s that provides maximum likelihood. All the samples taken out can be collated in a set defined as , where . The centroid of cluster χ s can be obtained by μ s = E[χ s]. It can be then employed to compute the distance of all the samples from μ s; i.e. d x = δ(x, μ s) ∀ x ∈ χ, where δ denotes a distance measure (in this case the Euclidean metric) and d x is a 1-dimensional sample or point corresponding to x. Thereafter, a centroid-based clustering scheme can be applied on this distance metric or inner product space (by considering 1-dimensional data and partitioning it into 2 groups) to readjust the cluster χ s. Here clustering is applied on a distance metric, which can be either the Euclidean norm or any form of kernel (as it is derived from dot product). This procedure can be repeated if μ s is changing dramatically. The overall method is summarized in Table 1 and illustrated in Additional file 1 (slides 1–6).
Table 1.
1. Given a sample set χ
s (at the beginning χ
s = χ), compute likelihood L (Eq. (12)). 2. Until L * > L, remove one sample (Eq. (18)), compute new likelihood L *, update χ s and L. 3. Find centroid μ s = E[χ s] and d x = δ(x, μ s) ∀ x ∈ χ. 4. Partition {d x} into two groups, for example using k-means algorithm (or divide into two groups based on their values). One of these groups will have lower d x values (representing closeness to μ s) whereas the other will have higher d x values (representing distance from μ s). Update χ s by replacing it with the samples with the lower d x values. 5. If required repeat steps 3 and 4. Take out the cluster χ s from χ. Update χ accordingly (the updated χ would contain all the samples except χ s; i.e. χ ∩ χ s = Φ). 6. Repeat all the steps until all the possible clusters (or desired number of clusters) are obtained. |
The next issue is the estimation of the number of clusters (c). If the value of c were given, it is then easier to find the locations. However, in some applications, c is unknown. In such situations, a range of values of c can be inserted to the procedure so that the best value in the range can be estimated. If no clue about c were given, the maximum number of possible clusters C can be investigated and the best among them (c ≤ C) can be chosen. In order to estimate c, we first define the total likelihood function as
19 |
where L i is the likelihood of ith cluster. The total likelihood function L t can be computed for different values of c. If for a particular number of clusters (k), the variation between two successive total likelihood functions were not significant, then we can estimate c to be k. Let the difference between two successive total likelihoods be δL t(k) = L t(k + 1) − L t(k), this quantity can be normalized as
20 |
where δL t is normalized over all the possible values of k.
Figure S5 in Additional file 1 illustrates the above explanation with a dataset of 4 clusters.
DRAGON method: search complexity
In this section, we briefly discuss the search complexity of the DRAGON method. As explained above the proposed method begins by taking one sample out of the sample set, which increases the likelihood function and requires n search. However, in the second iteration the search reduces to n − 1. Finding a cluster having n 1 samples requires (1/2)(n − n 1)(n + n 1 + 1) total search (see Additional file 2 for details). Therefore, the search for c clusters results O(n 2 c). It should be noted here that this search in conventional divisive hierarchical approaches is quite expensive, in the order of O(2n). However, the search of DRAGON is in the order of O(n 2 c), which indicates a considerable reduction. Furthermore, DRAGON employs the k-means clustering algorithm in the dot product space as an intermediate step. The computational complexity of k-means is considered to be linear (e.g. using Lloyd’s algorithm this is O(2nt) because dimensionality is 1 in an intermediate step, and the number of classes is 2. Here, t represents the number of iterations).
Results and discussion
To validate the DRAGON method, we performed analyses using synthetic and biological data. We further compared its clustering accuracy with that of existing hierarchical methods.
Analysis on synthetic data
For the analysis on synthetic data, we generated Gaussian data of dimensionality d with 4 clusters. This data consisted of 400 samples with similar topology to that described in Figure S1 of Additional file 1). With the help of different random seeds we produced the data 20 times, and for on each occasion we calculated the clustering accuracy. We then computed the average or mean of clustering accuracy over 20 attempts to have a statistically stable value. The dimension of the generated data was increased from 2 to 30. For evaluation purposes, we also used other agglomerative hierarchical methods such as SLink, CLink, MLink, ALink, Wa-Link, and Wt-Link. The average clustering accuracies of these approaches over dimensionality d are shown in Fig. 2a. For all the above methods, we provided the number of clusters; i.e. c = 4. For the DRAGON method (of Table 1), we iterated two times (step 5 of Table 1) in all the experiments. Additionally, we assessed seven previously compared divisive hierarchical methods [24]: ALink (average link), SLink (single link), CLink (complete link), DunnsOrg (Dunn’s original), DunnsVar (Dunn’s variant), Mac-Smith (Macnaughton-Smith) and PDDP (Principal Direction). The average clustering accuracies of these divisive methods are summarized in Fig. 2b. As Fig. 2b shows when the dimensionality of data increases, Mac-Smith performs poorly whereas ALink, SLink and DunnsOrg slightly improve their performances. The data dimension did not affect the accuracy of DunnsVar whatsoever, which remains around 50%. The highest accuracy is achieved by the PDDP method (roughly 80%), however, this accuracy is still lower than that of DRAGON.
It can be observed from Fig. 2 that on Gaussian data, the DRAGON method provides promising results over other hierarchical methods either agglomerative or divisive. The Wa-Link hierarchical method also shows good results after the DRAGON method over dimensionality d = 2, …, 30.
To avoid limiting our validation to Gaussian datasets, we carried out additional analyses on synthetic data that included Pathbased [48], Flame [49] and Aggregation [50] datasets. The results are summarized in Tables S1.1, S1.2 and S1.3 of Additional file 1.
Analysis on biological data
We also utilized biological datasets, namely acute leukemia [51], mixed-lineage leukemia (MLL) [52] and mutation data from The Cancer Genome Atlas for assessing clustering accuracy of several hierarchical approaches studied in this paper. A summary of datasets can be found below:
Acute leukemia dataset –comprises DNA microarray gene expressions. The samples belong to acute leukemias of humans. Two kinds are available: acute myeloid leukemia (AML) and acute lymphoblastic leukemia (ALL). The dataset has 25 AML and 47 ALL bone marrow samples with a dimension of 7129.
MLL dataset –has three ALL, MLL and AML classes. It contains 72 leukemia samples where 24 belong to ALL, 20 belong to MLL and 28 belong to AML with dimensionality 12,582.
Mutation dataset – this dataset is derived from The Cancer Genome Atlas project (https://tcga-data.nci.nih.gov/docs/publications/tcga/?). It includes mutation data for breast cancer, glioblastoma, kidney cancer and ovarian cancer. The data is divided into two groups of 416 samples and 90 samples, which contain 1636 genes.
To vary its number of features or dimensions, we employed the Chi-squared method for ranking the genes (the InfoGain feature selection method was also employed, see Additional file 3). We then performed clustering on samples and computed the clustering accuracy on dimensionality d = 2, …, 5. The clustering accuracies (was measured with the package AccMeasure 2011: http://www.mathworks.com/matlabcentral/fileexchange/32197-clustering-results-measurement/content/AccMeasure.m) on acute leukemia, MLL and mutation datasets are depicted in Tables 2, 3 and 4, respectively. The best outcomes are highlighted in bold faces.
Table 2.
Methods | Dim 2 | Dim 3 | Dim 4 | Dim 5 |
---|---|---|---|---|
SLINK | 66.7 | 66.7 | 66.7 | 66.7 |
CLINK | 84.7 | 81.9 | 81.9 | 81.9 |
ALINK | 76.4 | 81.9 | 84.7 | 84.7 |
Wa-LINK | 94.4 | 81.9 | 81.9 | 81.9 |
Wt-LINK | 94.4 | 81.9 | 81.9 | 81.9 |
MLINK | 94.4 | 81.9 | 81.9 | 81.9 |
SLINK (Divisive) | 66.7 | 66.7 | 66.7 | 66.7 |
CLINK (Divisive) | 80.6 | 80.6 | 80.6 | 80.6 |
ALINK (Divisive) | 66.7 | 66.7 | 66.7 | 66.7 |
Dunn’s original (Divisive) | 76.4 | 80.6 | 80.6 | 80.6 |
Dunn’s variant (Divisive) | 72.2 | 70.8 | 70.8 | 72.2 |
Macnaughton-Smith (Divisive) | 86.1 | 81.9 | 81.9 | 81.9 |
Principal Direction (Divisive) | 89.4 | 88.9 | 88.9 | 88.9 |
K-means | 90.3 | 89.5 | 81.9 | 81.9 |
DRAGON | 93.1 | 97.2 | 97.2 | 94.4 |
Table 3.
Methods | Dim 2 | Dim 3 | Dim 4 | Dim 5 |
---|---|---|---|---|
SLINK | 40.3 | 40.3 | 43.1 | 43.1 |
CLINK | 45.8 | 50.0 | 54.2 | 72.2 |
ALINK | 50.0 | 50.0 | 50.0 | 72.2 |
Wa-LINK | 62.5 | 62.5 | 62.5 | 84.7 |
Wt-LINK | 45.8 | 50.0 | 43.1 | 69.4 |
MLINK | 45.8 | 50.0 | 43.1 | 69.4 |
SLINK (Divisive) | 41.7 | 41.7 | 43.1 | 43.1 |
CLINK (Divisive) | 54.2 | 45.8 | 56.9 | 72.2 |
ALINK (Divisive) | 41.7 | 41.7 | 43.1 | 72.2 |
Dunn’s original (Divisive) | 44.4 | 44.4 | 45.8 | 72.2 |
Dunn’s variant (Divisive) | 41.7 | 41.7 | 43.1 | 73.6 |
Macnaughton-Smith (Divisive) | 54.2 | 48.6 | 50.0 | 72.2 |
Principal Direction (Divisive) | 62.5 | 62.5 | 62.5 | 81.9 |
K-means | 56.0 | 57.0 | 58.1 | 61.6 |
DRAGON | 65.3 | 62.5 | 68.1 | 84.7 |
Table 4.
Methods | Dim 2 | Dim 3 | Dim 4 | Dim 5 |
---|---|---|---|---|
SLINK | 77.9 | 77.9 | 77.9 | 82.4 |
CLINK | 77.3 | 82.0 | 82.8 | 83.2 |
ALINK | 77.3 | 77.3 | 82.8 | 82.4 |
Wa-LINK | 77.3 | 77.7 | 77.9 | 77.9 |
Wt-LINK | 77.3 | 82.0 | 82.8 | 54.6 |
MLINK | 54.6 | 54.6 | 54.6 | 82.4 |
SLINK (Divisive) | 54.5 | 54.5 | 54.5 | 82.2 |
CLINK (Divisive) | 54.4 | 54.5 | 77.3 | 77.3 |
ALINK (Divisive) | 77.9 | 82.6 | 82.6 | 82.4 |
Dunn’s original (Divisive) | 77.3 | 83.0 | 77.3 | 82.4 |
Dunn’s variant (Divisive) | 54.5 | 54.5 | 54.5 | 77.3 |
Macnaughton-Smith (Divisive) | 77.9 | 83.0 | 82.8 | 82.4 |
Principal Direction (Divisive) | 54.5 | 54.5 | 54.5 | 54.5 |
K-means | 64.9 | 67.1 | 65.7 | 63.3 |
DRAGON | 77.9 | 82.2 | 82.2 | 82.2 |
It is noticed from Table 2 that Wa-Link, Wt-Link and MLink achieve the highest performance when d = 2. For the other dimensions (d = 3, 4 and 5), DRAGON shows the highest performance. From Table 3, it can be seen that DRAGON achieves reasonable performance for all the dimensions. On mutation data (Table 4), DRAGON is showing the highest performance for d = 2 and slightly low performance (82.2%) for the other dimensions. Dunn’s original and Macnaughton-Smith provide the highest performance when d = 3. Despite the reasonable performance of divisive clustering methods in Table 4, it is worth noting that their running times were extremely slow, specially when the number of samples increases as it is the case with mutation data. In general, it can be summarized that the DRAGON method exhibited promising results in terms of clustering accuracy over other hierarchical methods. Also its search complexity is O(n 2 c), which is significantly lower than that of conventional divisive approaches.
Conclusions
In this work, we proposed a divisive hierarchical maximum likelihood clustering method whose search complexity was reduced to O(n 2 c). Its overall clustering accuracy showed a significant improvement over that of agglomerative hierarchical clustering methods when compared on both synthetic and biological datasets.
Additional files
Acknowledgements
We thank Maurice Roux for kindly providing us with his implementation of divisive hierarchical clustering algorithms. High-end computing resources were provided by the Advanced Center for Computing and Communication, RIKEN.
Funding
This study was funded by the CREST, JST Grant, Japan.
Availability of data and materials
The ALL and MLL datasets are publicly accessible and can be downloaded via the author’s webpage or the Kent Ridge Biomedical Data Set Repository. The mutation data can be retrieved from The Cancer Genome Atlas Project. The MATLAB package of DRAGON can be accessed at http://www.riken.jp/en/research/labs/ims/med_sci_math/ or http://www.alok-ai-lab.com
Declarations
The publication charges of this article were funded by JST CREST grant JPMJCR1412.
About this supplement
This article has been published as part of BMC Bioinformatics Volume 18 Supplement 16, 2017: 16th International Conference on Bioinformatics (InCoB 2017): Bioinformatics. The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume-18-supplement-16.
Abbreviations
- ALL
Acute lymphoblastic leukemia
- AML
Acute myeloid leukemia
- CLink
Complete linkage
- DIANA
Divisive analysis
- DRAGON
Divisive hierarchical maximum likelihood clustering
- EM
Expectation-maximization
- GMM
Gaussian mixture model
- ML
Maximum likelihood
- MLink
Median linkage
- MLL
Mixed lineage leukemia
- SLink
Single linkage
- Wa-Link
Ward’s linkage
- WLink
Weighted average distance linkage
- Wt-Link
Weighted linkage
Authors’ contributions
AS developed the concept, conducted the experiments and wrote the first draft of the manuscript. YL carried out the experiments related to divisive hierarchical clustering and proofread the manuscript. TT supervised the entire project. All the authors read and approved the final manuscript.
Ethics approval and consent to participate
None
Consent for publication
Not applicable
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Footnotes
Electronic supplementary material
The online version of this article (doi:10.1186/s12859-017-1965-5) contains supplementary material, which is available to authorized users.
Contributor Information
Alok Sharma, Email: alok.sharma@griffith.edu.au.
Yosvany López, Email: yosvany.lopez.alvarez@gmail.com.
Tatsuhiko Tsunoda, Email: tsunoda.mesm@mri.tmd.ac.jp.
References
- 1.Castro RM, Coates MJ, Nowak RD. Likelihood based hierarchical clustering. IEEE Trans Signal Process. 2004;52(8):2308–2321. doi: 10.1109/TSP.2004.831124. [DOI] [Google Scholar]
- 2.Defays D. An efficient algorithm for a complete link method. Comput J. 1977;20(4):364–366. doi: 10.1093/comjnl/20.4.364. [DOI] [Google Scholar]
- 3.Fisher D. Iterative optimization and simplification of hierarchical clusterings. J Artif Intell Res. 1996;4(1):147–179. [Google Scholar]
- 4.Goldberger J, Roweis S. Hierarchical clustering of a mixture model. In: Neural information processing systems. Cambridge: MIT Press; 2005. p. 505–12.
- 5.Heller KA, Ghahramani Z. Bayesian hierarchical clustering. In: 22nd international conference on machine learning. ACM: New York, Bonn; 2005. p. 297–304.
- 6.Horng S-J, M-Y S, Chen Y-H, Kao T-W, Chen R-J, Lai J-L, Perkasa CD. A novel intrusion detection system based on hierarchical clustering and support vector machines. Expert Syst Appl. 2011;38(1):306–313. doi: 10.1016/j.eswa.2010.06.066. [DOI] [Google Scholar]
- 7.Jain AK, Murty MN, Flynn PJ. Data clustering: a review. ACM Comput Surv. 1999;31(3):264–323. doi: 10.1145/331499.331504. [DOI] [Google Scholar]
- 8.Vaithyanathan S, Dom B. 16th conference in uncertainty in artificial intelligence. 2000. Model-based hierarchical clustering; pp. 599–608. [Google Scholar]
- 9.Sharma A, Boroevich KA, Shigemizu D, Kamatani Y, Kubo M, Tsunoda T. Hierarchical maximum likelihood clustering approach. IEEE Trans Biomed Eng. 2017;64(1):112–122. doi: 10.1109/TBME.2016.2542212. [DOI] [PubMed] [Google Scholar]
- 10.Zhang W, Zhao D, Wang X. Agglomerative clustering via maximum incremental path integral. Pattern Recogn. 2013;46(11):3056–3065. doi: 10.1016/j.patcog.2013.04.013. [DOI] [Google Scholar]
- 11.Rokach L, Maimon O. Clustering methods. In: Maimon O, Rokach L, editors. Data mining and knowledge discovery handbook. US: Springer; 2005. pp. 321–352. [Google Scholar]
- 12.Székely GJ, Rizzo ML. Hierarchical clustering via joint between-within distances: extending Ward’s minimum variance method. J Classif. 2005;22:151–183. doi: 10.1007/s00357-005-0012-9. [DOI] [Google Scholar]
- 13.Karypis G, Han E-H, Kumar V. Chameleon: hierarchical clustering using dynamic modeling. Computer. 1999;32(8):68–75. doi: 10.1109/2.781637. [DOI] [Google Scholar]
- 14.Murtagh F, Legendre P. Ward’s hierarchical agglomerative clustering method: which algorithms implement ward’s criterion? J Classif. 2014;31(3):274–295. doi: 10.1007/s00357-014-9161-z. [DOI] [Google Scholar]
- 15.Bouguettaya A, Yu Q, Liu X, Zhou X, Song A. Efficient agglomerative hierarchical clustering. Expert Syst Appl Int J. 2015;42(5):2785–2797. doi: 10.1016/j.eswa.2014.09.054. [DOI] [Google Scholar]
- 16.Zhang X, Xu Z. Hesitant fuzzy agglomerative hierarchical clustering algorithms. Int J Syst Sci. 2015;46(3):562–576. doi: 10.1080/00207721.2013.797037. [DOI] [Google Scholar]
- 17.Murtagh F, Contreras P. Algorithms for hierarchical clustering: an overview. WIREs Data Min Knowl Discov. 2012;2(1):86–97. doi: 10.1002/widm.53. [DOI] [Google Scholar]
- 18.Gómez D, Yáñez J, Guada C, Rodríguez JT, Montero J, Zarrazola E. Fuzzy image segmentation based upon hierarchical clustering. Knowl-Based Syst. 2015;87:26–37. doi: 10.1016/j.knosys.2015.07.017. [DOI] [Google Scholar]
- 19.Zhou G-T, Hwang SJ, Schmidt M, Sigal L, Mori G. Hierarchical maximum-margin clustering. Ithaca: Cornell University Library; 2015.
- 20.Wang J, Zhong J, Chen G, Li M, F-X W, Pan Y. ClusterViz: a cytoscape APP for cluster analysis of biological network. IEEE/ACM Trans Comput Biol Bioinform. 2015;12(4):815–822. doi: 10.1109/TCBB.2014.2361348. [DOI] [PubMed] [Google Scholar]
- 21.Kaufman L, Rousseeuw PJ. Finding groups in data: an introduction to cluster analysis. 1st ed: Wiley-Interscience; Hoboken, New Jersey; 2005.
- 22.Duda RO, Hart PE, Stork DG. Pattern classification. 2nd ed: Wiley-Interscience; New York; 2000.
- 23.Macnaughton-Smith P, T Williams W B. Dale M G. Mockett L: Dissimilarity analysis: a new technique of hierarchical sub-division. Nature 1964, 202:1034-1035. [DOI] [PubMed]
- 24.Roux M. A comparative study of divisive hierarchical clustering algorithms. Ithaca: Cornell University Library; 2015.
- 25.Roux M. Basic procedures in hierarchical cluster analysis. In: Devillers J, Karcher W, editors. Applied multivariate analysis in SAR and environmental studies. Dordrecht: Springer Netherlands; 1991. pp. 115–135. [Google Scholar]
- 26.Guénoche A, Hansen P, Jaumard B. Efficient algorithms for divisive hierarchical clustering with the diameter criterion. J Classif. 1991;8(1):5–30. doi: 10.1007/BF02616245. [DOI] [Google Scholar]
- 27.Farrell S, Ludwig CJH. Bayesian and maximum likelihood estimation of hierarchical response time models. Psychon Bull Rev. 2008;15(6):1209–1217. doi: 10.3758/PBR.15.6.1209. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Pinheiro J, Bates D. Mixed-effects models in S and S-PLUS. 2000th ed. New York: Springer; 2009.
- 29.Raudenbush SW, Bryk AS. Hierarchical linear models: applications and data analysis methods. 2nd ed. Thousand Oaks: SAGE Publications; 2001.
- 30.Sibson R. SLINK: an optimally efficient algorithm for the single-link cluster method. Comput J. 1973;16(1):30–34. doi: 10.1093/comjnl/16.1.30. [DOI] [Google Scholar]
- 31.Sokal RR, Michener CD. A statistical method for evaluating systematic relationships. Univ Kansas Sci Bull. 1958;38(22):1409–1438. [Google Scholar]
- 32.Legendre P, Legendre L. Numerical Ecology. 3rd ed: Elsevier; 2012.
- 33.Everitt BS, Landau S, Leese M, Stahl D. Cluster analysis. 5th ed: Wiley; 2011.
- 34.Ward JH. Hierarchical grouping to optimize an objective function. J Am Stat Assoc. 1963;58(301):236–244. doi: 10.1080/01621459.1963.10500845. [DOI] [Google Scholar]
- 35.Podani J. Multivariate data analysis in ecology and systematics: SPB Academic Publishing; 1994.
- 36.McQuitty LL. Similarity analysis by reciprocal pairs for discrete and continuous data. Educ Psychol Meas. 1966;26(4):825–831. doi: 10.1177/001316446602600402. [DOI] [Google Scholar]
- 37.McLachlan G, Krishnan T. The EM algorithm and extensions. 2nd ed: Wiley-Interscience; 2008.
- 38.Jordan MI, Jacobs RA. Hierarchical mixtures of experts and the EM algorithm. Neural Comput. 1994;6(2):181–214. doi: 10.1162/neco.1994.6.2.181. [DOI] [Google Scholar]
- 39.González M, Gutiérrez C, Martínez R. Expectation-maximization algorithm for determining natural selection of Y-linked genes through two-sex branching processes. J Comput Biol. 2012;19(9):1015–1026. doi: 10.1089/cmb.2010.0242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Do CB, Batzoglou S. What is the expectation maximization algorithm? Nat Biotechnol. 2008;26(8):897–899. doi: 10.1038/nbt1406. [DOI] [PubMed] [Google Scholar]
- 41.Holmes I. Using evolutionary expectation maximization to estimate indel rates. Bioinformatics. 2005;21(10):2294–2300. doi: 10.1093/bioinformatics/bti177. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Sharma A, Paliwal KK. Fast principal component analysis using fixed-point algorithm. Pattern Recogn Lett. 2007;28(10):1151–1155. doi: 10.1016/j.patrec.2007.01.012. [DOI] [Google Scholar]
- 43.Sharma A, Paliwal KK. A gradient linear discriminant analysis for small sample sized problem. Neural Process Lett. 2008;27(1):17–24. doi: 10.1007/s11063-007-9056-7. [DOI] [Google Scholar]
- 44.Sharma A, Paliwal KK. Cancer classification by gradient LDA technique using microarray gene expression data. Data Knowl Eng. 2008;66(2):338–347. doi: 10.1016/j.datak.2008.04.004. [DOI] [Google Scholar]
- 45.Sharma A, Paliwal KK. Rotational linear discriminant analysis technique for dimensionality reduction. IEEE Trans Knowl Data Eng. 2008;20(10):1336–1347. doi: 10.1109/TKDE.2008.101. [DOI] [Google Scholar]
- 46.Sharma A, Paliwal KK, Onwubolu GC. Class-dependent PCA, MDC and LDA: a combined classifier for pattern classification. Pattern Recogn. 2006;39(7):1215–1229. doi: 10.1016/j.patcog.2006.02.001. [DOI] [Google Scholar]
- 47.Sharma A, Paliwal KK. A new perspective to null linear discriminant analysis method and its fast implementation using random matrix multiplication with scatter matrices. Pattern Recogn. 2012;45(6):2205–2213. doi: 10.1016/j.patcog.2011.11.018. [DOI] [Google Scholar]
- 48.Chang H, Yeung DY. Robust path-based spectral clustering. Pattern Recogn. 2008;41(1):191–203. doi: 10.1016/j.patcog.2007.04.010. [DOI] [Google Scholar]
- 49.Fu LM, Medico E. FLAME, a novel fuzzy clustering method for the analysis of DNA microarray data. BMC Bioinformatics. 2007;83. pp. 1-15, https://doi.org/10.1186/1471-2105-8-3. [DOI] [PMC free article] [PubMed]
- 50.Gionis A, Mannila H, Tsaparas P. Clustering aggregation. ACM Trans Knowl Discov Data. 2007;1(1):1–30. doi: 10.1145/1217299.1217303. [DOI] [Google Scholar]
- 51.Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA, et al. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999;286(5439):531–537. doi: 10.1126/science.286.5439.531. [DOI] [PubMed] [Google Scholar]
- 52.Armstrong SA, Staunton JE, Silverman LB, Pieters R, den Boer ML, Minden MD, Sallan SE, Lander ES, Golub TR, Korsmeyer SJ. MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia. Nat Genet. 2002;30(1):41–47. doi: 10.1038/ng765. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The ALL and MLL datasets are publicly accessible and can be downloaded via the author’s webpage or the Kent Ridge Biomedical Data Set Repository. The mutation data can be retrieved from The Cancer Genome Atlas Project. The MATLAB package of DRAGON can be accessed at http://www.riken.jp/en/research/labs/ims/med_sci_math/ or http://www.alok-ai-lab.com