Skip to main content
FEBS Open Bio logoLink to FEBS Open Bio
. 2017 Nov 13;7(12):2008–2020. doi: 10.1002/2211-5463.12327

Multiple‐cumulative probabilities used to cluster and visualize transcriptomes

Xingang Jia 1,2,, Yisu Liu 3, Qiuhong Han 4, Zuhong Lu 2,
PMCID: PMC5715267  PMID: 29226087

Abstract

Analysis of gene expression data by clustering and visualizing played a central role in obtaining biological knowledge. Here, we used Pearson's correlation coefficient of multiple‐cumulative probabilities (PCC‐MCP) of genes to define the similarity of gene expression behaviors. To answer the challenge of the high‐dimensional MCPs, we used icc‐cluster, a clustering algorithm that obtained solutions by iterating clustering centers, with PCC‐MCP to group genes. We then used t‐statistic stochastic neighbor embedding (t‐SNE) of KC‐data to generate optimal maps for clusters of MCP (t‐SNE‐MCP‐O maps). From the analysis of several transcriptome data sets, we demonstrated clear advantages for using icc‐cluster with PCC‐MCP over commonly used clustering methods. t‐SNE‐MCP‐O was also shown to give clearly projecting boundaries for clusters of PCC‐MCP, which made the relationships between clusters easy to visualize and understand.

Keywords: icc‐cluster, multiple‐cumulative probabilities, PCC‐MCP, t‐SNE‐MCP


Abbreviations

ED‐MCP

Euclidean distance of multiple‐cumulative probabilities

FP of t‐SNE

the first components of t‐SNE projections

icc‐cluster

a clustering algorithm that obtains solutions by iterating clustering centers

MCP

multiple‐cumulative probability

PCA

principal component analysis

PCC‐MCP

Pearson's correlation coefficient of multiple‐cumulative probabilities

PCC

Pearson's correlation coefficient

SOM

self‐organizing map

SP of t‐SNE

the second components of t‐SNE projections

t‐SNE‐MCP‐k

t‐SNE of KC‐data

t‐SNE‐MCP‐O

the optimal t‐SNE‐MCP‐k map for clusters of MCP

t‐SNE‐MCP

t‐SNE of (n − 1)‐data

t‐SNE‐N

t‐SNE of the normalized points

t‐SNE

t‐statistic stochastic neighbor embedding

Clustering analysis is used to search for patterns and group genes into expression clusters that provide additional insight into the biological function and relevance of genes that show different expressions 1. The most popular clustering algorithms include hierarchical clustering 2, 3, k means clustering 4, and self‐organizing maps (SOMs) 5. However, clustering analysis cannot reveal underlying global patterns in the data, or relationships between the clusters found. To complement clustering analysis, dimension reduction techniques map the high‐dimensional points onto a 2D or 3D visualization space that is displayed graphically as a scatter plot, which provides a humanly interpretable visualization of the data set. A commonly used method for this purpose is principal component analysis (PCA). But for complex gene expression data sets, PCA typically gives poor visualizations 6, 7. Because of these limitations, non‐linear dimension reduction methods have been used to preserve local structure in the data, such as t‐statistic stochastic neighbor embedding (t‐SNE) 7, 8, 9, 10. t‐SNE has been successful in complementing clusters of Euclidean distance, but it is usually inefficient for displaying clusters of Pearson's correlation coefficient (PCC). For instance, for any reference data set in this paper, t‐SNE gives poor visualizations for the clusters that are generated by PCC.

Here, we use Pearson's correlation coefficient of multiple‐cumulative probabilities (PCC‐MCP) as a measure to define similarity of genes, where MCPs are composed by n cumulative probabilities of genes, and n is the dimension of gene points. These cumulative probabilities of each gene are generated from n permutations of the normalized points. For permutations of a normalized point, they have the same elements as the normalized points, but their element orders are different. Compared to the normalized points, MCPs are able to weaken the curve shape difference of genes with similar expression behavior. For instance, these genes can be seen as similar expression behavior that their elements are relatively equivalent, but their shape curves may have differences. Moreover, MCPs enlarge the element discrepancy of dissimilar genes. To evaluate the reliability of PCC‐MCP, we apply it with icc‐cluster to group a simulated data set and four experimental expression data sets. When PCC‐MCP with icccluster applies to these data sets, it produces clusters of more statistical relevance than those generated by some other popular clustering methods. This superior performance of PCC‐MCP partially confirms the validity of MCPs. Moreover, icc‐cluster with PCC‐MCP has great ability to remove the effect of the clustering numbers. In fact, even if clustering number is relative large or small compared to the optimal one, icccluster is able to attain tight and stable clusters.

Here, we firstly construct KC‐data sets, where KC‐data is composed of the first to k‐th principal components of MCPs, these components of MCPs are generated from PCA, k is less than n, and n is the dimension of the genes. Then, t‐SNE‐MCP‐k map is generated from t‐SNE of KC‐data, where t‐SNE‐MCP‐k is 2D projections of the KC‐data. And then, t‐SNE‐MCP‐O map is selected from these t‐SNE‐MCP‐k projections by the average silhouette value of clustering results. That is, t‐SNE‐MCP‐O is such t‐SNEMCP‐k map that has the most clear boundaries for 2D projections of clustering results. To evaluate the reliability of t‐SNE‐MCP‐O, we use it to display clusters of a simulated data set and four experimental expression data sets, where clusters are generated from PCC‐MCP. Results show that t‐SNE‐MCP‐O gives clearer projecting boundaries for these clusters than commonly used dimension‐reduction techniques. Furthermore, to readily see which nearby 2D points are truly similar, we also construct gene neighbor maps by t‐SNE‐MCP‐(n – 1). Results show that t‐SNE‐MCP‐(n – 1) makes the relationships between clusters easy to visualize and understand.

In this study, icc‐cluster with PCC‐MCP is firstly applied to yeast metabolic cycle data 11, mouse retina data 12, human embryo data 13 and K562 cell line data 14, and then these clustering results are overlayed onto t‐SNE‐MCP‐O maps, which makes the identification of gene relationships easy and intuitive.

Materials and methods

Data set 1

The simulation data set consisted of 2000 members at five time points, which were generated from normal distributions. These 2000 members belonged to four groups, A, B, C and D, according to the models they were generated from, and each group consisted of 500 members. The members of groups that were generated from the different normal distributions are shown in Table 1.

Table 1.

List of simulated data

Time points Group A Group B Group C Group D
1 N(10,2) N(10,2) N(100,10) N(10,2)
2 N(10,2) N(0.5,0.1) N(60,6) N(30,3)
3 N(10,2) N(0.5,0.1) N(60,6) N(10,2)
4 N(10,2) N(0.5,0.1) N(60,6) N(60,6)
5 N(10,2) N(30,3) N(100,10) N(10,2)

Data set 2

This data set consisted of yeast metabolic cycle data: NCBI GEO accession number GSE3431. It described the transcriptional changes in the metabolic cycle of budding yeast, Saccharomyces cerevisiae 11. In this experiment, gene expression behaved in a periodic manner, comprising a non‐respiratory phase followed by a respiratory phase. The transcriptome was assayed every 25 min over three consecutive cycles, resulting in 36 samples (T1–T36). These were profiled using Affymetrix YG_S98 oligonucleotide arrays (Affymetrix Inc., Santa Clara, CA, USA). Probes that had at least three present calls generated by Affymetrix gene chip software were classified as expressed and the data normalized using genespring v7 per‐chip normalization. Using a periodicity algorithm described in the original paper, the authors classified 3552 genes as periodic, corresponding to 3656 probe sets. From these 3656 probe sets, 2913 probes, expression values were greater than 5 in at least one of 36 samples selected. These 2913 probes were summarized in Data S2.

Data set 3

The raw mouse retinal data consisted of 10 SAGE libraries (38 818 unique tags with tag counts ≥ 2) from developing retina taken at 2‐day intervals. The samples ranged from embryonic, to postnatal, to adult. Among the 38 818 tags, 1467 tags that had counts ≥ 20 in at least one of the 10 libraries were selected 12. The purpose of this selection was to exclude the genes with uniform low expression. The counts of each tag in a SAGE library were Poisson distributed. These Poisson distributions were independent of each other across different tags and libraries 12, 15. These 1467 tags are summarized in Data S3.

Data set 4

This data set consisted of human embryo data: NCBI GEO accession number GSE18887. The resulting matrix contained expression measurements for 5441 transcripts across 18 samples, denoted as the human organogenesis expression matrix 13 (Carnegie stages 9–14, S9–S14). A total of 5441 probe sets were identified as differentially expressed using extraction of differential gene expression (EDGE)‐based methodology. Initially, Fang et al. used SOM‐SVD (SOM combined with singular value decomposition) to identify co‐expressed genes of human embryo data 13, which identified six clusters. From their analysis, they extracted 2148 differentially expressed probe sets. We used this set of 2148 probe sets for our analysis. These 2148 probes are summarized in Data S4.

Data set 5

NCBI GEO accession number GSE12736. Time course microarray data were obtained at seven independent time points. Duplicate experiments were performed for each time point. Selecting genes with a significant detection P‐value produced 14 000 probes out of a total of 23 920 probes. Quantile normalization was carried out for each data set at seven time points using the average expression value. It was reasoned that significant genes should show over two‐fold induction at least at one time point with respect to the control sample (t = 0; before phorbol 12‐myristate 13‐acetate treatment), and 1779 probes satisfying this requirement had been determined 14, 16. These 1779 probes are summarized in Data S5.

MCPs of genes

Here, X i = {x i1, x i2, ···, x in} represents the i‐th gene, and x ij represents the expression level of the j‐th time points. Here, we sketch MCP of X i as follows.

  • (a)

    X i is normalized into Y i, where

Yi={yi1,yi2,,yin},yit=xitmin(min1tn(xit),0)l=1n(xilmin(min1tn(xit),0)),t=1,2,,n.

For genes, expression levels may be negative at some time points, such as the genes of data set 5. Here, we deal with these genes by a translational transformation. That is, x it is substituted by xitmin(min1tn(xit),0). In fact, if all expression levels of X i are non‐negative, xitmin(min1tn(xit),0) is the same as x it. Moreover, Z is is constructed, where Z is is the s‐th permutation of Y i, and

Zis={yis,yi(s+1),,yin,yi1,yi2,,yi(s1)},s=1,2,,n.

Based on Z is, T is is constructed, where

Tis={yis/2,yis+yi(s+1)/2,,l=sn1yil+yin/2,l=snyil+yi1/2,l=snyil+yi1+yi2/2,,l=snyil+l=1s2yil+yi(s1)/2}. (1)

That is, T is is the modified cumulative probability of Z is.

  • (b)

    Based on T is, T i is constructed, where T i is an n 2‐dimensional vector, and

Ti={Ti1,Ti2,,Tin}.

Here, T i is considered to be the MCP of X i.

To clearly understand MCPs, we used a flowchart to show their construction process (Fig. 1). Moreover, we define Euclidean distance and PCC between T i and T j as the Euclidean distance of multiple‐cumulative probabilities (ED‐MCP) and PCC‐MCP of X i and X j, respectively.

Figure 1.

Figure 1

Flowchart of the multiple‐cumulative probabilities.

icc‐cluster algorithm

  • (a)

    Choose X j1 and X j2 as the first and second cluster centers, where

dj1,j2=max1i<jm{di,j},

m is the gene number of the data set. If X js satisfies

min{djs,j1,djs,j2,,djs,j(s1)}=max1im,ij1,,j(s1)min{di,j1,di,j2,,di,j(s1)},

it is chosen as the s (s = 4, 5, ···, k)‐th cluster center.

  • (b)

    X i belongs to the l‐th cluster (1 ≤ lk) if it satisfies

di,jl=min{di,j1,di,j2,,di,jk}.
  • (c)

    Assuming X l1, X l2, ···, X lq in the l‐th cluster, X li is chosen as a new center if it satisfies

j=1qdli,lj=min1mqj=1qdlm,lj.
  • (d)

    Repeat step 2 and step 3 until the assignment does not change.

For the icc‐cluster algorithm, when PCC is used to measure the similarity of genes,

di,j=1φi,j,

where φi,j is PCC between X i and X j (or T i and T j).

The t‐SNE‐MCP‐O and t‐SNE‐MCP‐k maps

The formal description of the t‐SNE algorithm can be found in Ref. 9. But for the high‐dimensional expression points of genes, it is difficult for us to obtain t‐SNE projections of their MCPs. To answer the challenge of the high dimensional MCPs, we use t‐SNE‐MCP‐O to display clusters of PCC‐MCP, where t‐SNE‐MCP‐O selects from t‐SNE‐MCP‐k (k = 2, 3, ···, n − 1) maps, t‐SNE‐MCP‐k is t‐SNE of KC‐data, points of KC‐data are composed of the first to k‐th principal components of T i, components of T i were generated from PCA. In fact, for PCA of MCPs, only their first to (n − 1)‐th principal components are not zero, and we abbreviate t‐SNE‐MCP‐(n − 1) as t‐SNE‐MCP.

Here, for clusters of PCC‐MCP, we use the average silhouette value of t‐SNE‐MCP‐k to select their t‐SNE‐MCP‐O. The average silhouette value of t‐SNE‐MCP‐k is defined as

Sk=1mi=1m(biai)max(ai,bi),

where a i is the average Euclidean distance from U i to the other points in the same cluster as U i, b i is the minimum average distance from U i to points in a different cluster, minimized over clusters, U i is the t‐SNE projection of P i, P i is the i‐th point of KC‐data, and m is the gene number of the data 17. Moreover, if T i belongs to the a‐th cluster, cluster membership of U i is a also. For these S k of clusters that generate from PCC‐MCP, if S k is the largest, t‐SNEMCP‐k is selected as the t‐SNE‐MCP‐O map of the clusters. For convenience, we use t‐SNE‐MCP‐1 to denote t‐SNE of MCPs.

Results

Here, all clustering results were generated from the normalized points, and PCC‐MCP, PCC, ED‐MCP, Euclidean distance, TransChisq and PoissonC were chosen as measures of genes. Moreover, clustering number of data sets mainly came from the corresponding references. In detail, data set 2 had been divided into eight clusters by Euclidean distance 7, data set 3 into 25 clusters by PoissonC and TransChisq 12, 15, data set 4 into six and 10 clusters by Euclidean distance 7, and data set 5 into eight clusters by Euclidean distance 14.

Comparison of PCC‐MCP and PCC

Here, data set 1 was firstly displayed on t‐SNE‐MCP‐O and t‐SNE‐N maps according to population membership of points (Fig. 2A,B), where t‐SNE‐MCP‐O of data set 1 was t‐SNE‐MCP‐3, and population membership of points was summarized in Table 1. From Fig. 2A,B, t‐SNE‐MCP‐O and t‐SNE‐N correctly separated members of different groups, and only a few members of group A were assigned to group C, where members of group A and group C (Table 1) were marked by the blue and red colors in Fig. 2A,B, respectively. Moreover, data set 1 was divided into four clusters by PCC‐MCP and PCC, where icc‐cluster and k means were used simultaneously. Then, these clustering results were displayed on t‐SNE‐MCP‐O and t‐SNE‐N maps (Fig. 2C–F).

Figure 2.

Figure 2

Overlay of clusters of data set 1 onto 2D maps, where data points were colored according to cluster membership. (A) Overlay of four populations onto t‐SNE‐MCP‐O map. (B) Overlay of four populations onto t‐SNE‐N map. (C) Overlay of four clusters onto t‐SNE‐MCP‐O map, where clusters were generated from icc‐cluster with PCC‐MCP. (D) Overlay of four clusters onto t‐SNE‐N map, where clusters were generated from icc‐cluster with PCC. (E) Overlay of four clusters onto t‐SNE‐MCP‐O map, where clusters were generated from k means with PCC‐MCP. (F) Overlay of four clusters onto t‐SNE‐N map, where clusters were generated from k means with PCC.

From Fig. 2A,C,E, for clusters of PCC‐MCP, icc‐cluster and k means correctly assigned members of group B, C and D into three different clusters, and only a few members of group A were assigned to the cluster that contained all the points of group C. But for clusters of PCC, icc‐cluster and k means assigned members of group A to four different clusters, and assigned members of group C to three different clusters (Fig. 2B,D,F). For members of group A, their elements of five time points came from N(10,2) simultaneously, so their elements were relatively equivalent, but their shape curves had significant differences. The poor performance of PCC may be due to the fact that PCC only compared the shape of the curves. For instance, PCC between a1 = (8.08, 12.36, 9.86, 10.63, 11.14) and a4 = (12.39, 9.91, 9.54, 9.35, 9.41) was only −0.74, while PCC between a1 and d384 = (6.49, 37.18, 9.12, 49.39, 13.12) was 0.594, where a1 and a2 came from group A: and d384 came from group D. However, PCC‐MCP between a1 and a4 was 0.993, while PCC‐MCP between a1 and d384 was 0.95. That is, the defect of PCC was removed by MCPs.

The statistical reliability of icc‐cluster with PCC‐MCP

The average silhouette value was a quantitative way to evaluate the clustering solutions 17, and we used it to demonstrate the reliability of icc‐cluster with PCC‐MCP. Here, we applied ED‐MCP, PCC‐MCP, PCC and Euclidean distance to four experimental data sets. For comparison, icc‐cluster and k means were used simultaneously, and the average silhouette values of these clustering results were summarized in Table 2. For all data sets, Table 2 showed that their largest average silhouette value came from icc‐cluster, and three of these largest average silhouette values were generated by PCC‐MCP, where the largest average silhouette value of each data set was marked by an asterisk in Table 2. Moreover, for clustering results of any data, the average silhouette value of ED‐MCP was far less than PCC‐MCP. That is, clusters of PCC‐MCP were better separated from neighboring clusters than ED‐MCP.

Table 2.

The average silhouette values of different measures. Number: clustering number. For each data set, the largest average silhouette value was marked by an asterisk

Data Algorithm Number ED‐MCP PCC‐MCP Euclidean distance PCC
2 icc‐cluster 3 0.24905 0.40003 0.48845* 0.47263
2 k means 3 0.25236 0.40542 0.26024 0.47327
3 icc‐cluster 25 0.18948 0.31151* 0.13419 0.22494
3 k means 25 0.18638 0.30046 0.15225 0.23674
4 icc‐cluster 10 0.32985 0.53427* 0.27882 0.20489
4 k means 10 0.14063 0.31636 0.16123 0.11007
5 icc‐cluster 12 0.26081 0.41184* 0.23601 0.30403
5 k means 12 0.16611 0.2457 0.17662 0.38694

The biological reliability of icc‐cluster with PCC‐MCP

Here, 34 ‘cell‐specific’ tags of data set 3 were used to test the biological reliability of icc‐cluster with PCC‐MCP, where these 34 tags are summarized in Data S3. Moreover, these ‘cell‐specific’ tags showed the most dynamic and cell‐specific expression in the mouse neonatal retina (developmental stages P 0P 6) 12. Data set 3 had been grouped into 25 clusters by TransChisq and PoissonC measure 12, 15, respectively. Moreover, these 34 cell‐specific genes were used to demonstrate that TransChisq and PoissonC measures were more efficient for analyzing SAGE data than PCC and Euclidean distance. For comparison, we used icc‐cluster with PCC‐MCP to group these 1467 tags into 25 clusters also. Then, for each of the different algorithms, the five most dynamic clusters that contained ‘cell‐specific’ tags were selected. The comparison statistics of 34 ‘cell‐specific’ tags were summarized in Table 3. In Table 3, icc‐cluster generated clusters that were most enriched for these 34 cell‐specific genes. That is, PCC‐MCP was appropriate and reliable for analyzing SAGE data also.

Table 3.

Statistics of 34 ‘specific’ genes of data set 3. The second column showed the numbers of cell‐specific genes in a cluster; total, the total number of cluster members; sensitivity, the number of cell‐specific genes/34; precision, the number of cell‐specific genes/total number of cluster members. The top five clusters that contained the 34 cell‐specific genes are listed. The numbers in bold were the highest percentage for sensitivity and precision for that method

Cell‐specific genes Total Sensitivity (%) Precision (%)
PCC‐MCP 11 41 32.3 26.8
1 5 2.8 20
2 14 5.9 14.3
5 38 14.7 13.2
3 27 8.8 11.1
TransChisq 9 41 26.5 22.0
1 12 2.9 8.3
3 39 8.8 7.7
1 24 2.9 4.2
3 74 8.8 4.1
PoissonC 10 47 29.4 21.3
4 22 11.8 18.2
3 27 8.8 11.1
1 11 2.9 9.1
1 12 2.9 8.3

The features of the MCPs

By PCC‐MCP with icc‐cluster, data set 5 was firstly divided into eight clusters. Then, these clusters were selected to explore the feature of MCPs. For MCPs and normalized points of each cluster, their curve shapes were shown in Fig. 3. For the curve shapes of MCPs, Fig. 3 showed that they were almost the same in the same clusters, but had significant differences in the different clusters. That is, MCPs weakened the curve shape difference of the genes with similar expression behavior, but enlarged the element discrepancy of dissimilar genes. Importantly, for each of clusters that were generated from PCC‐MCP, the curve shapes of their normalized points had no significant difference (Fig. 3A–H).

Figure 3.

Figure 3

The profile plots of the normalized points and MCPs. The X‐axis represented the different time points. The Y‐axis represented the expression level, where E‐level is the abbreviation of expression level. (A–H) The profiles of normalized plots of eight clusters; (I–P) The profile of the multiple‐cumulative probabilities of eight clusters.

The features of icc‐cluster

Here, we used data set 2 to validate that icc‐cluster had great ability to remove the effect of clustering numbers, where data set 2 was divided into three and five clusters by PCC‐MCP. The comparison statistics of super‐clusters and sub‐clusters are summarized in Table 4. For each sub‐cluster of icc‐cluster, Table 4 showed that it contained 85% genes that came from the same super‐cluster at least. However, for two sub‐clusters of k means, their genes came from the same super‐cluster that were < 70%.

Table 4.

Statistics of three superclusters and five subclusters of data set 2. The second column shows the number of subclusters. The numbers in the third, fourth and fifth column were the number of genes in subclusters that came from the three superclusters. Total was the total number of subcluster members. The sensitivity was the bold numbers/total number of subcluster members

No. 1 2 3 Total Sensitivity (%)
icc‐cluster 1 16 0 0 16 100.00
2 257 0 0 257 100.00
3 0 0 1458 1458 100.00
4 14 996 106 1116 89.25
5 65 11 0 76 85.53
k means 1 431 4 220 655 65.80
2 0 496 0 496 100.00
3 0 0 271 271 100.00
4 828 393 0 1221 67.81
5 293 0 17 310 94.52

Moreover, we used t‐SNE‐MCP‐O (t‐SNE‐MCP‐2) maps to display the features of icc‐cluster (Fig. 4). For sub‐clusters of icc‐cluster, Fig. 4A,C showed that they mainly came from the same super‐clusters. From Fig. 4B,D, clustering results of k means were significantly affected by clustering numbers.

Figure 4.

Figure 4

Overlay of clusters of data set 2 onto t‐SNE‐MCP‐O (t‐SNE‐MCP‐2) maps, where data points were colored according to cluster membership. (A) Overlay of three superclusters of icc‐cluster onto t‐SNE‐MCP‐O map. (B) Overlay of three superclusters of k means onto t‐SNE‐MCP‐O map. (C) Overlay of five subclusters of of icc‐cluster onto t‐SNE‐MCP‐O map. (D) Overlay of five subclusters of k means onto t‐SNE‐MCP‐O map.

The consistency between clusters of PCC‐MCP and t‐SNE‐MCP

For a visualization technique, if it was able to project genes of the same clusters together, and project neighbor clusters in adjacent regions, we considered that it was consistent with clustering results. In general, we assessed the consistency by eye, which was an intuitive feeling only. Here, the average silhouette value was used to quantify the consistency, where we focused on data set 4 and 5. Firstly, data set 4 was divided into 10 clusters by PCC‐MCP (or ED‐MCP) with icc‐cluster (or k means). For any clustering result of data set 4, the average silhouette values of its t‐SNE‐MCP‐k maps were shown in Fig. 5A. From Fig. 5A, for t‐SNE‐MCP, the average silhouette values of any clustering result were less than t‐SNE‐MCP‐15. That is, t‐SNE‐MCP‐15 was t‐SNE‐MCP‐O of data set 4, and t‐SNEMCP‐1 did not optimally display clusters of MCPs. Secondly, data set 5 was divided into 8, 12, 16 and 20 clusters by PCC‐MCP with icc‐cluster. For any clustering result of data set 4, the average silhouette values of its t‐SNE‐MCP‐k maps were shown in Fig. 5B. From Fig. 5B, for 8, 16 and 20 clusters, t‐SNE‐MCP‐2 was their t‐SNE‐MCP‐O. But for 12 clusters, the average silhouette value t‐SNE‐MCP‐1 was the largest. That is, for different clustering number, their t‐SNE‐MCP‐O maps were not consistent.

Figure 5.

Figure 5

The average silhouette values of t‐SNE‐MCP‐k and D 1‐plot of data set 4 and 5. (A) The average silhouette values of t‐SNE‐MCP‐k maps of data set 4, where clustering results of PCC‐MCP with icc‐cluster, ED‐MCP with icc‐cluster, PCC‐MCP with k means and ED‐MCP with k means were shown by red, green, blue and gray line, respectively. (B) The average silhouette values of all t‐SNE‐MCP‐k maps of data set 5, where 8, 12, 16 and 20 clusters were shown by red, green, blue and gray line, respectively. (C) D 1‐plot of data set 4, where D 1‐plot of t‐SNE‐MCP, t‐SNE‐MCP‐O, PCA‐MCP, PCA‐CP and PCA‐N were shown by blue, pink, green, gray and red line, respectively. (D) D 1‐plot of data set 5, where D 1‐plot of t‐SNE‐MCP, t‐SNE‐MCP‐O, PCA‐MCP, PCA‐CP and PCA‐N were shown by blue, pink, green, gray and red line, respectively.

The local validity of t‐SNE‐MCP

We considered that a dimension‐reduction technique was locally valid if the i‐th closest neighbour of a point was its j‐th closest neighbour in 2D space, where i, j and |ij| were the relative small numbers, point neighbours were located by any measure, while projection neighbours were located by Euclidean distance 6. Moreover, the local validity of dimension reduction techniques could be quantified by D 1‐plot 6.

Here, data set 4 and 5 were used to assess the local validity of t‐SNE‐MCP, t‐SNE‐MCP‐O, PCA‐MCP, PCA‐CP and PCA‐N, where we named the PCA of the normalized points and cumulative probabilities as PCA‐N and PCA‐CP, respectively. Moreover, when gene neighbours were defined by PCC of MCPs, the normalized points and cumulative probabilities, their neighboring maps corresponded to t‐SNE‐MCP (t‐SNE‐MCP‐O and PCA‐MCP), PCA‐CP and PCA‐N, respectively. Moreover, D 1‐plots of data set 4 and 5 are shown in Fig. 5C and D, respectively. From Fig. 5C,D, the local validity of t‐SNE‐MCP was far more than others. That is, t‐SNE‐MCP better preserved local proximities of genes compared with other dimension‐reduction techniques.

Comparison of t‐SNE‐MCP and t‐SNE‐MCP‐O

By PCC‐MCP with icc‐cluster, data sets 4 and 5 were firstly divided into 10 and 8 clusters, respectively. Then, these clustering results were displayed on t‐SNE‐MCP and t‐SNE‐MCP‐O maps (Fig. 6). As shown in Fig. 6A,B, t‐SNE‐MCP‐O and t‐SNE‐MCP gave good projections for clusters of data set 4, where t‐SNE‐MCP‐O was t‐SNE‐MCP‐15. For clusters of data set 5, t‐SNE‐MCP‐O (t‐SNE‐MCP‐2) gave clearly projecting boundaries (Fig. 6C), while their t‐SNE‐MCP map showed slight intermixing (Fig. 6D). In fact, for data sets 2 and 3, t‐SNE‐MCP maps were not the optimal projections for clusters of PCC‐MCP also.

Figure 6.

Figure 6

Overlay of clusters of data set 4 and 5 onto t‐SNE maps, where clusters were generated from PCC‐MCP with icc‐cluster, and data points were colored according to cluster membership. (A) Overlay of 10 clusters of data set 4 onto t‐SNE‐MCP‐O (t‐SNE‐MCP‐15) map. (B) Overlay of 10 clusters of data set 4 onto t‐SNE‐MCP map. (C) Overlay of eight clusters of data set 5 onto t‐SNE‐MCP‐O (t‐SNE‐MCP‐2) map. (D) Overlay of eight clusters of data set 5 onto t‐SNE‐MCP map.

Comparison of t‐SNE‐MCP‐O, t‐SNE‐N and t‐SNE‐C

Here, data set 4 was divided into six clusters by PCC‐MCP (or PCC) with k means, and data set 2 into three clusters by PCC with k means. Moreover, these clustering results were displayed on t‐SNE‐MCP‐O, t‐SNE‐N and t‐SNE‐C maps (Fig. 7), where t‐SNE‐C was t‐SNE of the centered gene points 7. From Fig. 7, only t‐SNE‐MCP‐O gave clearly projecting boundaries for clusters of PCC‐MCP (Fig. 7A), while t‐SNE‐N and t‐SNE‐C maps had significant intermixing for any clustering result of PCC (Fig. 7B,C,D).

Figure 7.

Figure 7

Overlay of clusters of data set 4 and 2 onto t‐SNE‐MCP‐O, t‐SNE‐N and t‐SNE‐C maps, where data points were colored according to cluster membership. (A) Overlay of six clusters of data set 4 onto t‐SNE‐MCP‐O (t‐SNE‐MCP‐15) map, where clusters were generated from PCC‐MCP with k means. (B) Overlay of six clusters of data set 4 onto t‐SNE‐N map, where clusters were generated from PCC with k means. (C) Overlay of three clusters of data set 2 onto t‐SNE‐C map, where clusters were generated from PCC with k means. (D) Overlay of three clusters of data set 2 onto t‐SNE‐N map, where clusters were generated from PCC with k means.

Comparison of ED‐MCP and Euclidean distance

By ED‐MCP (or ED‐N) with k means, data set 4 was divided into six clusters, and data set 5 into eight clusters, where ED‐N was Euclidean distance of the normalized gene points. Moreover, clustering results of data set 4 were displayed on t‐SNE‐MCP maps (Fig. 8A,B), and clustering results of data set 5 on t‐SNE‐N maps (Fig. 8C,D). From Fig. 8, only t‐SNE‐MCP gave clearly projecting boundaries for clusters of ED‐MCP (Fig. 8B), and t‐SNE‐N maps had slight intermixing. Importantly, Fig. 8 showed that clustering results of ED‐MCP and ED‐N had no significant difference. That is, MCPs retained the difference of the normalized genes.

Figure 8.

Figure 8

Overlay of clusters of data set 4 and 5 onto t‐SNE‐MCP and t‐SNE‐N maps, where data points were colored according to cluster membership. (A) Overlay of six clusters of data set 4 onto t‐SNE‐MCP map, where clusters were generated from ED‐N with k means. (B) Overlay of six clusters of data set 4 onto t‐SNE‐MCP map, where clusters were generated from ED‐MCP with k means. (C) Overlay of eight clusters of data set 5 onto t‐SNE‐N map, where clusters were generated from ED‐N with k means. (D) Overlay of eight clusters of data set 5 onto t‐SNE‐N map, where clusters were generated from ED‐MCP with k means.

The gene neighbors of t‐SNE‐MCP

Hierarchical clustering was commonly used to reveal gene neighbors, for it was much faster and more memory‐efficient 2, 3. However, hierarchical clustering was likely to cause loose gene neighbors. That is, two neighbors that were generated by the hierarchical clustering were not really the nearest and second‐closest neighbors of genes 7.

Here, the nearest and second‐closest gene neighbors were generated by PCC‐MCP, where we focused our attention on data set 4. For genes of data set 4, their neighbor maps were shown on a t‐SNE‐MCP map in Fig. 9, where the nearest gene neighbors were linked by a blue line, and second‐closest gene neighbors were linked by a red line. Compared with the hierarchical clustering, the gene neighbor map revealed the pairs of high‐dimensional points that were truly close, and which pairs were in fact distant in 2D space. Moreover, t‐SNE‐MCP maps combined with nearest neighbour maps provided an intuitive means to understand the relationship between clusters and the affiliation of genes with specific clusters.

Figure 9.

Figure 9

The gene neighbors of data set 4. The nearest and second closest neighbors of genes of data set 4, where the nearest gene neighbors were linked by a blue line, and second‐closest gene neighbors were linked by a red line.

Discussion

Although the cumulative probabilities have one‐to‐one mapping with their normalized points, their magnitudes have significant differences, which may distort distance of some similar genes. Moreover, for the different position elements of a normalized point, their superposed opportunity is not consistent in cumulative probability, which can distort the similarity of genes. The defect of cumulative probabilities is removed by MCPs. That is, for MCPs, the magnitudes are the same, and the superposed opportunities of elements in normalized points are consistent.

For high‐throughput data sets, it is difficult for us to obtain t‐SNE projections of the MCPs. Here, for clusters of PCC‐MCP, we use t‐SNE‐MCP‐O to obtain their optimal 2D maps. Moreover, icc‐cluster can relatively rapidly achieve convergence to the optimal solutions. The reason is that icc‐cluster only updates the cluster centers.

Conclusions

The success of MCPs has two main aspects. One is that MCPs remove the differences of the curve shape of similar expression genes, which makes PCC‐MCP able to robustly measure the similarity of genes. Another is that MCPs enlarge the element divagations of dissimilar genes, which make t‐SNE‐MCP‐O able to clearly display clustering results of PCC‐MCP. We suggest that MCPs can provide new insights applicable to analyzing high‐throughput data. Furthermore, MATLAB implemented PCC‐MCP with icccluster, and from Figs 2–9 are available at Data S1.

Author contributions

XJ analyzed and discussed the model, and wrote the manuscript. QH and YL performed a portion of the model. ZL supervised the study.

Supporting information

Data S1. A freely available matlab algorithm implemented to obtain clusters from PCC‐MCP with icc‐cluster, and used in Figs 2–9.

Data S2. Data set 2 (yeast metabolic cycle data: NCBI GEO accession number GSE3431). Its detail (see Ref. [7]).

Data S3. Data set 3 (the raw mouse retinal data consisted of 10 SAGE libraries from developing retina taken at 2‐day intervals). Its detail (see Ref. [12]).

Data S4. Data set 4 (human embryo data: NCBI GEO accession number GSE18887). Its detail (see Ref. [7]).

Data S5. Data set 5 (NCBI GEO accession number GSE12736). Its detail (see Ref. [14]).

Acknowledgements

This work rests almost entirely on open data. Contributors are gratefully acknowledged. This work is supported by Major Program of National Natural Science Foundation of China (2016YFA0501600).

Contributor Information

Xingang Jia, Email: hanqh15@163.com.

Zuhong Lu, Email: zhlu@seu.edu.cn.

References

  • 1. Gehlenborg N, O'Donoghue SI, Baliga NS, Goesmann A, Hibbs MA, Kitano H, Kohlbacher O, Neuweger H, Schneider R, Tenenbaum D et al (2010) Visualization of omics data for systems biology. Nat Methods 7, s56–s68. [DOI] [PubMed] [Google Scholar]
  • 2. Eisen MB, Spellman PT, Brown PO and Botstein D (1998) Cluster analysis and display of genome‐wide expression patterns. Proc Natl Acad Sci USA 95, 14863–14868. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Wen X, Fuhrman S, Michaels GS, Carr DB, Smith S, Barker JL and Somogyi R (1998) Large‐scale temporal gene expression mapping of central nervous system development. Proc Natl Acad Sci USA 95, 334–339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Tavazoie S, Hughes JD, Campbell MJ, Cho RJ and Church GM (1999) Systematic determination of genetic network architecture. Nat Genet 22, 281–285. [DOI] [PubMed] [Google Scholar]
  • 5. Tamayo P, Slonim D, Mesirov J, Zhu Q, Kitareewan S, Dmitrovsky E, Lander ES and Golub TR (1999) Interpreting patterns of gene expression with self‐organizing maps: methods and application to hematopoietic differentiation. Proc Natl Acad Sci USA 96, 2907–2912. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Jia X, Zhu G, Han Q and Lu Z (2017) The biological knowledge discovery by PCCF measure and PCA‐F projection. PLoS One 12, e0175104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Natascha B, James S, James B and Christopher W (2011) An intuitive graphical visualization technique for the interrogation of transcriptome data. Nucleic Acids Res 39, 1–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Van der ML and Hinton G (2008) Visualizing data using t‐SNE. J Mach Learn Res 9, 2579–2605. [Google Scholar]
  • 9. Venna J and Kaski S (2006) Local multidimensional scaling. Neural Netw 19, 889–899. [DOI] [PubMed] [Google Scholar]
  • 10. Taskesen E and Reinders MJ (2016) 2D representation of transcriptomes by t‐SNE exposes relatedness between human tissues. PLoS One 11, e0149853. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Tu BP, Kudlicki A, Rowicka M and McKnight SL (2005) Logic of the yeast metabolic cycle: temporal compartmentalization of cellular processes. Science 310, 1152–1158. [DOI] [PubMed] [Google Scholar]
  • 12. Li C, Haiyan H, Seth B, Jun SL, Connie C and Wing HW (2004) Clustering analysis of SAGE data using a Poisson approach. Genome Biol 5, R51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Fang H, Yang Y, Li C, Fu S, Yang Z, Jin G, Wang K, Zhang J and Jin Y (2010) Transcriptome analysis of early organogenesis in human embryos. Dev Cell 19, 174–184. [DOI] [PubMed] [Google Scholar]
  • 14. Limb JK, Yoon S, Lee KE, Kim BH, Lee S, Bae YS, Jhon GJ and Kim J (2009) Regulation of megakaryocytic differentiation of K562 cells by FosB, a member of the Fos family AP‐11 transcription factors. Cell Mol Life Sci 66, 1962–1973. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Kim K, Zhang S, Jiang K, Cai L, Lee IB, Feldman LJ and Huang H (2007) Measuring similarities between gene expression profiles through new data transformations. BMC Bioinformatics 8, 29. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Saeed AI, Bhagabati NK, Braisted JC, Liang W, Sharov V, Howe EA, Li J, Thiagarajan M, White JA and Quackenbush J (2006) TM4 microarray software suite. Methods Enzymol 411, 134–193. [DOI] [PubMed] [Google Scholar]
  • 17. Rousseeuw PJ (1987) Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J Comput Appl Math 20, 53–65. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data S1. A freely available matlab algorithm implemented to obtain clusters from PCC‐MCP with icc‐cluster, and used in Figs 2–9.

Data S2. Data set 2 (yeast metabolic cycle data: NCBI GEO accession number GSE3431). Its detail (see Ref. [7]).

Data S3. Data set 3 (the raw mouse retinal data consisted of 10 SAGE libraries from developing retina taken at 2‐day intervals). Its detail (see Ref. [12]).

Data S4. Data set 4 (human embryo data: NCBI GEO accession number GSE18887). Its detail (see Ref. [7]).

Data S5. Data set 5 (NCBI GEO accession number GSE12736). Its detail (see Ref. [14]).


Articles from FEBS Open Bio are provided here courtesy of Wiley

RESOURCES