Abstract
We introduce a principled approach for simultaneous mapping and clustering (SMAC) for establishing consistent maps across heterogeneous object collections (e.g., 2D images or 3D shapes). Our approach takes as input a heterogeneous object collection and a set of maps computed between some pairs of objects, and outputs a homogeneous object clustering together with a new set of maps possessing optimal intra- and inter-cluster consistency. Our approach is based on the spectral decomposition of a data matrix storing all pairwise maps in its blocks. We additionally provide tight theoretical guarantees for the accuracy of SMAC under established noise models. We also demonstrate the usefulness of our approach on synthetic and real datasets.
1. Introduction
Establishing maps (e.g. pointwise correspondences) across object collections is a fundamental problem spanning many scientific domains. High-quality maps facilitating information propagation and transformation are key to applications ranging from 3D reconstruction with partial scans (Huber & Hebert, 2001), data-driven geometry completion and reconstruction (Pauly et al., 2005), texture transfer (Schreiner et al., 2004; Kraevoy & Sheffer, 2004), to comparative biology (Boyer et al., 2011; Gao et al., 2017), joint data- analysis (Huang et al., 2011; Kim et al., 2012; Wang et al., 2013; 2014; Huang et al., 2014), and data exploration and organization (Kim et al., 2012; Huang et al., 2014).
High quality object maps are generally difficult to compute. Prior work on map computation focused on optimizing maps between pairs of objects; see (van Kaick et al., 2010) for a standard survey and (Kim et al., 2011; Mandad et al., 2017) for some recent advances. Despite the significant progress, state-of-the-art techniques tends to hit a barrier on the quality of maps that are computed in a pairwise manner. Building upon the availability of big-data, a recent line of research (Kim et al., 2012; Nguyen et al., 2011; Ovsjanikov et al., 2012; Huang et al., 2012; Huang & Guibas, 2013; Huang et al., 2014; Chen et al., 2014; Zhou et al., 2015a;b; Shen et al., 2016; Leonardos et al., 2017; Huang et al., 2017) considered computing many pairwise maps jointly among a collection of objects. The promise of these approaches hinges upon the observation that one way to obtain a high quality pairwise map between dissimilar objects is to choose a path connecting these objects but consisting of consecutive similar shapes: maps between similar objects are typically of higher quality, and so is the resulted composed map. From a regularization perspective, joint map computation leverages the generic consistency of a network of maps among multiple objects, in which composition of maps along cycles are expected to be close to the identity map.
However, the performance of these data-driven approaches relies predominantly on the homogeneity of the object collection, i.e. the input objects fall into the same category or sub-category (e.g. Chairs, Cars, and Human models). In the presence of heterogeneous data, where the input objects fall into multiple underlying categories, applying existing data-driven approaches without the category label information tend to produce unsatisfactory results. In this setting, even though existing methods are able to suppress the noise in intra-cluster maps within a single cluster, jointly computed maps for the entire object collection leads are often significantly worse. One explanation is that high fraction of incorrect inter-cluster maps tends to “contaminate” the regularization effect of intra-cluster maps. A natural resolution is to employ a two-stage cascadic strategy that identifies the underlying clusters before computing the intra- and inter-cluster maps. Unfortunately, such clustering requires accurate quantification of the object similarities, which is a difficult problem in its own right. Meanwhile, the error produced in the clustering stage is unlikely remedied by the consistency-based regularization.
In this paper, we propose to solve the mapping and clustering problems simultaneously. Instead of explicitly relying on certain pairwise similarity and/or map distortion scores, we identify the underlying clusters based on the consistency of intra- and inter-cluster maps, inspired by the observation that maps tend to be more consistent along cycles within a cluster than across clusters. This discrepancy has been observed in many different contexts, and appears to be a consequence of the energy landscape of almost all optimization-based pairwise matching algorithms. The matching energy functional between objects in the same underlying cluster tends to have simple energy landscapes with easily identifiable global optimums, resulting in fairly cycle-consistent intra-cluster maps; in contrast, the highly non-convex energy landscape for dissimilar objects from different clusters leads to more “random” maps due to random initialization and/or multiple strong local minimums, for which cycle-consistency is much less often observed. This map consistency argument is the foundation of our simultaneous mapping and clustering (SMAC) algorithm.
1.1. Motivating Example
We validate the map consistency argument through a motivating example (see Figure 1) on a real dataset from SHREC07 Watertight benchmark (Giorgi et al., 2007). This dataset consists of 38 shapes: the first 18 are Human models and the remaining 20 are Fourleg models (e.g., Horses and Dogs). Each shape is represented as a discrete metric space with 1024 sample points generated from farthest-point sampling (Eldar et al., 1997). We compute pairwise blended intrinsic maps (Kim et al., 2011) for all objects in this collection and use these maps to compute two similarity scores for each object pair: a map distortion score that measures the squared sum of geodesic distortions across all point-pairs (c.f. (Bronstein et al., 2006)), and a cycle-consistency score that is the median value (among other options) of the distortion scores of all 3-cycles to which the pair belongs, where the distortion of a 3-cycle is defined as the squared sum of the geodesic distances between each point and its image propagated along the 3-cycle.
Figure 1.
The pairwise map distortion scores (Left) and the pair-wise cycle-consistency scores (Right) derived from 3-cycles. The scores are normalized by the maximum score in each matrix.
Figure 1 illustrates the distributions of the map distortion scores (Left) and the cycle-consistency scores (Right) on the 38 models. The cycle-consistency scores clearly reveal the underlying cluster structure and in fact better separates the two clusters of models (Human vs. Fourleg) than the map-distortion scores (intra-cluster blocks in the right figure are darker in blue). The superior cluster separation is verified by comparing the results of spectral clustering (Ng et al., 2002; Lei & Rinaldo, 2015) using the two similarity scores: spectral clustering based on the cycle-consistency scores recovers the two underlying clusters perfectly, whereas the same procedure using the map distortion scores incorrectly puts two Fourleg models in the cluster of Human models. This motivating example illustrates the effectiveness and superiority of the map consistency score as a quantification of object similarity.
1.2. Approach Overview
Motivated by the example above, we propose an algorithm for simultaneous mapping and clustering (SMAC). Our SMAC algorithm takes as input (i) an object collection that falls into multiple clusters, and (ii) some noisy maps pre-computed between object pairs, and outputs the underlying clusters together with improved maps between all pairs of objects. Our SMAC algorithm builds upon the equivalence between map consistency and the low-rank property of a data matrix with consistent maps in its blocks (c.f. (Huang & Guibas, 2013)). We show that this low-rank property still holds in the setting of multiple disjoint collections of consistent intra-cluster maps, though the rank is expected to be higher due to multiple clusters. Based on this observation, the first step of our approach simultaneously recovers the underlying clusters and intra-cluster maps by spectral decomposition. We show that properly “rounding off” the leading eigenvectors recovers the ground-truth clusters and intra-cluster maps in a single pass. We then construct inter-cluster maps from the recovered intra-cluster maps. Our theoretical analysis establishes sharp exact recovery conditions for both steps under a fairly general noise model, using some novel tight L∞-stability bounds for eigen-decompositions.
1.3. Related Works
Joint object matching, i.e., simultaneously estimating maps among a collection of objects, is an emerging field across many scientific problems. Earlier works use combinatorial optimizations (Nguyen et al., 2011; Kim et al., 2012; Huang et al., 2012). More recent works (Huang & Guibas, 2013; Huang et al., 2014; Chen et al., 2014; Wang & Singer, 2013) rely on convex or non-convex optimization techniques.However, all these methods assume that the underlying object collection is homogeneous (all objects belong to a single category). For a heterogeneous object collection where the objects fall into multiple distinctive clusters, existing methods usually rarely succeed in producing both high-quality intra- and inter-cluster maps.
Clustering and in particular spectral clustering is another well-studied topic. We refer to (Lei & Rinaldo, 2015; Rohe et al., 2011) for some recent advances and to (Filippone et al., 2008; Luxburg, 2007; Fortunato, 2010) for surveys. Our approach falls into the general category of graph-based clustering, but the pairwise information we utilize is of “functional” rather than “scalar” nature. Instead of the more common approach that derives affinity scores from the pair-wise maps for clustering, our SMAC algorithm discovers the cluster structure based purely on the consistency of pairwise maps and demonstrates improved empirical performance. This strategy is reminiscent of heterogeneous multireference alignment (Boumal et al., 2017) and simultaneous alignment and classification (Lederman & Singer, 2016) for synchronization problems in Cryo-Electron Microscopy; in this context, our residue-based clustering strategy is closest in nature to learning group actions (Gao et al., 2016).
Our approach relies on tight L∞-type bounds on leading eigenvectors of perturbed matrices. Though the stability of eigenvalues and eigenspaces are well-studied, element-wise eigenvector stability appears to be a much harder problem; see recent survey (O’Rourke et al., 2016). We introduce new stability bounds to tackle this technical difficulty.
1.4. Mathematical Notation
We use lower bold letters a, b, c, u, v, w, . . .to denote vectors, and upper letters A, B, C, . . .for matrices. For a block matrix to de- note its ij-th block; the ij-th element of a matrix A is denoted as aij. With ⊗ we denote the Kronecker product. For a symmetric matrix we always sort the eigen-values in non-increasing order, i.e., Matrix norms will be used for a matrix is the maximum singular value of A, and the spectral norm is sometimes simplified as for the set of permutation matrices of dimension m × m.
2. Algorithm
For simplicity, we focus on describing and analyzing our algorithm under the setting where pair-wise maps are given by permutations. In Section 4, we show how to modify the algorithm for partially similar objects.
We first describe the problem setup. Consider n objects each represented by m points. With we denote an observation graph among An initial map is pre-computed on each edge using an off-the-shelf pairwise object matching algorithm from Si to Sj. We also assume the objects in are partitioned into k ≥ 2 clusters, but k is unknown. Our goal is to identify the underlying clusters, and in the mean-
Algorithm 1.
PermSMAC: Simultaneously mapping and clustering
Input: Observation graph and initial pairwise maps |
Output: Underlying clusters and optimized pairwise maps |
1: {Step 1} Simultaneously compute the intra-cluster maps and extract the underlying clusters: |
2: {Step 1.1} Form data matrix based on (1). |
3: {Step 1.2} Compute the critical value |
4: {Step 1.3} Let store the leading r eigen- vectors of X. Compute pair-wise maps by solving (2) |
5: {Step 1.4} Use as the affinity score and apply single-linkage clustering to obtain the underlying clusters |
6: {Step 2} compute the inter-cluster maps by solving (6) |
while improve all pairwise maps between objects in . As a basis for identifying the underlying clusters, we assume that the intra-cluster maps are more accurate (in terms of cycle-consistency) than inter-cluster maps.
Our algorithm proceeds in two major steps. The first step simultaneously extracts the underlying clusters and computes intra-cluster maps. The second step then computes inter-cluster maps. Now we introduce these two steps in details. Algorithm 1 provides the pseudo code.
2.1. Step I: Extract underlying clusters and compute intra-cluster maps.
Motivated from prior works for map synchronization (Pachauri et al., 2013; Shen et al., 2016), we use a block data matrix to encode the input maps:
(1) |
Our approach is motivated by the empirical observation that the leading eigen-vectors of X reveal the underlying clusters and simultaneously denoise intra-cluster maps if intra-cluster maps are more accurate than inter-cluster maps. We provide the algorithmic details below; an analysis is provided in Section 3.
Given the data matrix, we first estimates the number of stable eigen-vectors as
Here we search within the range of [m, nm], as we expect multiple underlying clusters. Let store the top r eigen-vectors of X, and divide U into n matrices U1, . . . , Un of shape m × r such that We then compute the estimated map along each edge by solving a linear assignment problem, i.e.,
(2) |
for all Note that (2) admits an exact solution via linear assignment. In fact,
(3) |
(4) |
(5) |
Intuitively, provides an approximation of the underlying map Xij, and the linear assignment procedure projects this approximation onto the space of permutations.
For clustering, we treat the residual score as the distance measure between and apply single-linkage clustering (Gower & Ross, 1969) to obtain the underlying clusters. We set as the number of desired clusters.
Empirically, we found that when the input inter-cluster maps are inaccurate, the quality of estimated inter-cluster maps appear to be much more noisy than estimated intra-cluster maps. This motivates us to re-estimate inter-cluster maps as a second step, described in Section 2.2 below.
2.2. Step II: Estimate inter-cluster maps.
We estimate the inter-cluster maps between each pair of clusters independently. Specifically, consider two clusters cs and ct . Let be a pair of objects selected from each cluster, respectively. We optimize the inter-cluster map represented as a pairwise object map by solving the following linear assignment:
(6) |
Note that it is possible to jointly optimize among all pairs of clusters. However, since the number of clusters is usually significantly smaller than the size of each cluster, we found the gain of doing so is insignificant.
3. Analysis
We first describe our noise model in Section 3.1. We then analyze our method under this model and present the exact recovery conditions for both underlying clusters and the pairwise maps in Section 3.2. Our analysis is based on a set of new stability bounds of eigen-decompositions. The proofs of all Lemma’s and Theorem’s with technical details can be referred to in the supplemental material.
3.1. Model for Analysis
We consider two models, one for the pairwise map and the other one for the observation graph and the underlying cluster structure.
Map model.
We generalize the map model described in (Shen et al., 2016) to multiple clusters: Suppose there are k underlying clusters. With cs , 1 ≤ s ≤ k we denote the vertex indices of the s-th cluster. In other words, we have Given an observation graph, the input pairwise maps are independent, and they follow
where denotes a random permutation matrix satisfying
(7) |
ηij depends on the edge type: ηij = p if (i, j) is an intra-cluster edge, i.e., and ηij = q if (i, j) is an inter-cluster edge, i.e., We assume p > q.
Remark 1.
Note that one can also generalize the model by assuming there exist underlying permutations so that the ground-truth map Nevertheless, it turns out that the two models are identical (Shen et al., 2016) . Hence we adopt the simpler form for convenience.
Model for the observation graph and clusters.
To obtain concise and interpretable exact recovery conditions, we are particularly interested in analyzing our algorithm when the observation graph and underlying clusters are generated from the following established model: assume and the size of each underlying cluster the observation graph is generated from the Erdὄs-Rényi G(n; t) with edge selection probability t. However, the stability bounds we introduce in the supplemental material can be used for analyzing more general noisy models, e.g., sizes of the underlying clusters are different.
3.2. Map and Cluster Recovery
We begin by analyzing the leading eigenvalues and eigen-vectors of E[X] to gain insights on the necessary conditions for map and cluster recovery. To make our discussion more general, we first assume the underlying cluster and the observation graph are fixed. Consider then the following (p, q)- reweighted normalized adjacency matrix
(8) |
It is clear that
and thus the non-zero eigenvalues of E[X] are non-zero eigenvalues of with multiplicity m−1. Furthermore, let be an orthonormal basis for Then the leading k(m − 1) eigenvectors of E[X] are given by This leads two conditions on the eigenvalues and eigenvectors of for map and cluster recovery:
Eigenvalue separation. Since our method leverages the largest eigengap, we assume that has the largest eigengap. Define
Then a necessary condition for map recovery is γ < 1.
Eigenvector separation. We further assume that the underlying clusters can be recovered by reading off the rows of Sk. Formally, consider rows of Sk as coordinates of the corresponding objects, and define
(9) |
(10) |
dintra and dinter essentially measure the maximum distance within each cluster and the minimum distance between different clusters, respectively. Thus, a necessary condition for cluster recovery is that dintra < μ dinter for some small constant μ.
Under these two conditions, it is easy to see that when X ≈ E[X], we have It follows that both the underlying clusters and intra-cluster maps can be exactly recovered.
These two separation conditions are quite general. In fact, they hold for the noisy model described in Section 3.1. To gain some further insight, one can show that (c.f. (Le et al., 2017)) with high probability
indicating that Moreover, under this model
If p and q are well-separated, the top k eigenvectors of approximate Now we formally state the exact recovery conditions:
Theorem 3.1.
(Intra-Cluster and Map Recovery) Assume Consider the noise model described in Section 3.1. There exists an absolute constant cintra such that PermSMAC recovers the underlying intra-cluster maps and the underlying clusters with high probability if
(11) |
Remark 2.
Note that the gap between p and q is used to ensure the recovery of the underlying clusters. Moreover, the recovery rate matches the information theoretic lower bound established in (Chen et al., 2016) up to indicating the tightness of our condition for PermSMAC.
The following theorem provides an exact recovery condition for the inter-cluster maps. Compared with the previous lower bound, the lower bound on q for inter-cluster map recovery is significantly cruder. This shows the advantage of recovering inter-cluster maps as a separate step.
Theorem 3.2.
There exists an absolute constant cinter > 0, so that when
PermSMAC recovers the underlying inter-cluster maps with high probability.
4. Partial Matching
In this section we extend the algorithm to handle partially (as opposed to fully) similar objects. Each input object in this setting has points, where the mi’s vary across the collection. Consequently, the pairwise maps are no longer permutation matrices since some rows and columns may only contain zero elements. We propose the following modified algorithm for SMAC in this partial matching setting.
Step I: Extract underlying clusters and compute intracluster maps.
Forming the data matrix and leading eigen-vector computation stay the same as in the full matching setting, except we replace in (1). The first difference occurring in the partial matching setting is that we cannot apply (2) to obtain pair-wise maps and affinity scores for clustering. Our strategy is to apply the distance measure between two points p, p′ is given by where up is the corresponding row of p in U . The number of output clusters is set as r. Each output cluster of this single-linkage procedure collects a set of matched points among the input objects. We merge two clusters if the objects they belong to overlap. Suppose we finally obtain k clusters For each cluster ci, we introduce a binary matrix whose columns encode the enclosed point clusters. Then it is easy to see that the blocks of describe intra-cluster maps. Note that ri = m in the full setting, but in the partial setting ri is usually larger than due to partial similarity.
Step II: Compute inter-cluster maps.
In the partial setting, we encode the inter-cluster map from cluster cs to cluster ct as a matrix Consider a object pair With Ei,s and Ej,t we denote the index matrices that extract the corresponding blocks in Ys and Yt. It is easy to see that the entries provide cues for the inter-cluster map Similar to the full map setting, we compute
Since the inter-cluster maps may not be permutation matrices either, we apply a simple thresholding to obtain the inter-cluster maps:
where βst is set as 0.9 times the maximum element of Cst in our experiments.
5. Experimental Results
In this section, we evaluate our approach on both synthetic (Section 5.1) and real datasets (Section 5.2 and Section 5.3). For baseline comparison, we consider state-of-the-art approaches for clustering and joint mapping in each domain.
5.1. Experimental Results on Synthetic Instances
We apply the model described in Section 3 to generate the synthetic data sets for our experiments. Below we summarize how the procedure depends on the model parameters:
The observation graph We employ a standard two- community stochastic block model (Abbe et al., 2016) which enables us to fully control the vertex degree and the spectral gap. We use this model to generate three observation graphs All of them have n = 300 vertices, but the vertex degrees and spectral gaps vary. Specifically, is the clique graph. is a sparse graph, whose average vertex degree is 50 and the spectral gap is 0.1. The average vertex degree and spectral gap of are 50 and 0.5, respectively. In our experiment, we treat as the default observation graph. We also study the influence of the observation graphs on the performance of our algorithm.
Number of clusters k. Without loss of generality, we allocate each object into a underlying cluster with probability For each observation graph, we generate and fix one underlying cluster throughout our experiments.
Other parameters m, γ, p, q. We fix the number of points on each object as m = 30. In the partial matching setting, we follow the protocol (Chen et al., 2014) to generate the input objects so that the expected size of each object is mγ. We sample the ratio of correct inter-cluster maps Since p > q, we sample the ratio of correct intra-cluster maps so that
We now study the empirical phase transition curves when varying p and q under different k, γ and observation graphs.
Varying k.
The first two rows of Figure 2 show the phase transition curves of map and cluster recovery for k = 2, 4, 6. Across all configurations, our approach tolerates a significant portion of noise in the input maps. The fraction of noise that our approach can handle reduces as k increases, which is consistent with the exact recovery conditions in Section 3. In addition, phase transitions with respect to mapping recovery and cluster recovery roughly align. The subtle differences are two-fold: when p and q are close, cluster recovery breaks as there is no cue for clustering; likewise, map recovery breaks when q approaches 0.
Figure 2.
Phase transition curves when varying q and p − q under different configurations of k. The first and second rows show mapping and cluster recovery of PermSMAC, respectively. The third row shows mapping recovery of (Huang & Guibas, 2013). (a) k = 2. (b) k = 4. (c) k = 6. Red means more success.
Versus mapping only.
Figure 2 compares our approach to state-of-the-art map recovery technique (Huang & Guibas, 2013). SMAC clearly exhibits a clear advantage in the regime, where q is small and p is significantly larger than q. This is expected through our exact recovery condition.
Varying observation graph.
Figure 3 shows phase transition curves of map and cluster recovery when varying the observation graphs (k = 2, γ = 1). Our approach tolerates a larger fraction of incorrect maps for larger vertex degrees. Moreover, when vertex degrees are comparable, a small spectral gap means higher recovery rate.
Figure 3.
Phase transition curves of map recovery and cluster recovery when varying the observation graph. The top row shows mapping recovery, and the bottom row shows cluster recovery.
Varying γ.
Figure 4 shows the phase transition curves when varying the overlapping ratio γ. We again show three configurations, i.e., γ = 1, γ = 0.8 and γ = 0.6. Still, our approach can tolerate a large rate of noise in the input maps. Moreover, the rate reduces as γ becomes smaller. This is expected, as low overlapping ratio introduces weak signal for mapping and cluster recovery.
Figure 4.
Phase transition curves of map recovery and cluster recovery when varying the overlapping ratio. The top row shows mapping recovery, and the bottom row shows cluster recovery. (a)γ = 1. (b) γ = 0.8. (c)γ = 0.6.
5.2. Experimental Results on 3D Shapes
We proceed to evaluate SMAC on 3D shapes. We consider two datasets for this task. The first dataset collects four categories of 3D shapes from SHREC07-Watertight (Giorgi et al., 2007), namely, Human, Fourleg, Armadillo and Teddy. These categories appear to have both similar global structures and local geometric details. However, the intercategory variability is salient. The second dataset is more fine-grained. It has 10 underlying shape collections from FAUST training dataset (Bogo et al., 2014), where each collection consists of different poses of the same human subject (10 poses per collection). For evaluating shape maps, we follow the protocol of (Kim et al., 2011) by collecting statistics on the geodesic distortion of predicted correspondences with respect to human annotated feature correspondences.
Mapping performance.
Figure 5(c) plots the accuracy of predicted correspondences of our approach versus the input. For a detailed assessment, we separate the statistics of intra-cluster maps and inter-cluster maps. We consider two approaches for baseline comparison: the first applies (Huang & Guibas, 2013) to the entire dataset, and the second applies (Huang & Guibas, 2013) to each category in isolation then applies the third step of our approach to compute inter-cluster maps. The second baseline may be considered as a performance upper bound. Our proposed SMAC is significantly better than mapping without clustering (which is seriously affected by the noise in inter-cluster maps) and competitive against the second baseline.
Figure 5.
Experimental results on 3D shapes. We consider two datasets: (left) Human/Fourleg/Armadillo/Teddy, and (right) Different poses of xxx human subjects. For each dataset, we show (top) snapshots of each dataset, and (bottom) error distribution of predicted intra-cluster and inter-cluster shape maps.
Clustering performance.
As shown in Table 1, our approach correctly identifies all underlying clusters in SHREC07 and FAUST. We also applied two baseline clustering approaches on the same dataset. The first approach performs k-means on the spectral shape descriptors (Rustamov, 2007). This approach only yields 84.6% and 72.0%, respectively. The second approach utilizes the mapping distortion as an affinity measure and applies spectral clustering. This approach yields 94.9% and 74.0%, respectively, which are better than the first baseline. However, our approach is still better, which shows the advantage of using the cycle-consistency constraint for clustering.
Table 1.
Classification accuracy of our method and baseline approaches on the four datasets show in Figure 5 and Figure 6. Distort. and Descrip. stand for using the pairwise mapping distortion score and global object descriptor for clustering, respectively.
SHREC07 | FAUST | Notre Dame | Church | |
---|---|---|---|---|
SMAC | 100% | 94.00% | 99.30% | 96.10% |
Distort. | 94.90% | 74.00% | 94.30% | 91.90% |
Descrip. | 84.60% | 72.00% | 84.30% | 86.70% |
5.3. Experimental Results on 2D Images
Finally, we evaluate our approach on two datasets of 2D images. The first dataset (Figure 6(Left)) consists of 600 internet images of Notre Dame. These images naturally fall into 3 categories, each of which collects images from a similar camera pose (Snavely et al., 2006). The second dataset (Figure 6(Right)) collects 600 internet images of 4 landmark churches in Europe (Amiens Cathedral (200 images), York Minster (200 images), Duomo (100 images) and Westminster Abbey (100 images)). As inter-cluster maps do not make much sense here, we only evaluate clustering results and intra-cluster maps in this experiment. We sample 400 SIFT features (Lowe, 2004) for each image and apply SIFT flow (Liu et al., 2011) to establish pairwise correspondences between the features. We manually mark feature correspondences for evaluation.
Figure 6.
Experimental results on 2D images. We consider two datasets: (Left) Internet images of Norte Dame from three diverse camera poses, and (Right) Internet images of five churches from similar view points. Visualization is the same as Figure 5.
Mapping performance.
Figure 6 compares our approach with the two baseline approaches introduced in Section 5.2. The relative performance is consistent. Specifically, due to the small-overlapping region across different clusters, inter-cluster maps are rather noisy, so applying joint-mapping directly leads to sub-optimal results. In addition, our approach is competitive against the approach of computing intra-cluster and inter-cluster maps in a sequential manner.
Clustering performance.
Finally, we evaluate our approach in terms of clustering accuracy. We choose two baseline approaches, where the first baseline approach performs k-means on image descriptors. In this case, we employ GIST (Oliva & Torralba, 2001). The second baseline uses the residual of SIFT flow as the affinity score for clustering. As shown in Table 1, our approach leads to a clustering accuracy of 99.3% and 96.1% on Notre Dame and Church, respectively. They are higher than those of the top performing baselines, i.e., 94.5% and 92.1%, respectively.
6. Conclusions
We have introduced SMAC for simultaneously computing consistent maps across a heterogeneous data collection and identifying the underlying clusters. The key idea is to leverage the higher self-consistency within intra-cluster maps than inter-cluster maps. Enforcing this variation of consistency allows us to denoise the input maps in a sequential manner while simultaneously identifying the underlying cluster structures. The approach is based on spectral decomposition, for which we provided tight exact recovery conditions for both the input maps and the underling clusters. Experimental results on synthetic data sets justify our exact recovery conditions, and experimental results on real data sets demonstrate the efficacy of our approach.
Supplementary Material
Acknowledgement.
This research by Chandrajit Bajaj was supported in part by a grant from the NIH, R01 GM117594. Tingran Gao acknowledges support from Simons Math+X Investigators Award 400837, DARPA D15AP00109, and NSF IIS 1546413. Qixing Huang would like to acknowledge support for this research from NSF DMS-1700234, NSF CIP-1729486, and NSF IIS-1618648.
References
- Abbe E, Bandeira AS, and Hall G Exact recovery in the stochastic block model. IEEE Transactions on Information Theory, 62(1):471–487, 2016. [Google Scholar]
- Bogo F, Romero J, Loper M, and Black MJ FAUST: Dataset and evaluation for 3D mesh registration. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), Piscataway, NJ, USA, June 2014. IEEE. [Google Scholar]
- Boumal N, Bendory T, Lederman RR, and Singer A Heterogeneous multireference alignment: a single pass approach. arXiv preprint arXiv:1710.02590, 2017. [Google Scholar]
- Boyer DM, Lipman Y, St. Clair E, Puente J, Patel BA, Funkhouser T, Jernvall J, and Daubechies I Algorithms to Automatically Quantify the Geometric Similarity of Anatomical Surfaces. Proceedings of the National Academy of Sciences, 108(45):18221–18226, 2011. doi: 10.1073/pnas.1112822108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bronstein AM, Bronstein MM, and Kimmel R Generalized multidimensional scaling: A framework for isometry-invariant partial surface matching. In PNAS, volume 103, pp. 1168–1172, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y, Guibas LJ, and Huang Q Near-optimal joint object matching via convex relaxation. In ICML, pp. 100–108, 2014. [Google Scholar]
- Chen Y, Suh C, and Goldsmith AJ Information recovery from pairwise measurements. IEEE Trans. Information Theory, 62(10):5881–5905, 2016. [Google Scholar]
- Eldar Y, Lindenbaum M, Porat M, and Zeevi YY The farthest point strategy for progressive image sampling. IEEE Trans. Image Processing, 6(9):1305–1315, 1997. [DOI] [PubMed] [Google Scholar]
- Filippone M, Camastra F, Masulli F, and Rovetta S A survey of kernel and spectral methods for clustering. Pattern Recogn., 41(1):176–190, January 2008. [Google Scholar]
- Fortunato S Community detection in graphs. Physics Reports, 486(3–5):75–174, 2010. [Google Scholar]
- Gao T, Brodzki J, and Mukherjee S The Geometry of Synchronization Problems and Learning Group Actions. arXiv preprint arXiv:1610.09051, 2016. [Google Scholar]
- Gao T, Yapuncich GS, Daubechies I, Mukherjee S, and Boyer DM Development and Assessment of Fully Automated and Globally Transitive Geometric Morphometric Methods, With Application to a Biological Comparative Dataset With High Interspecific Variation. The Anatomical Record, pp. n/a–n/a, 2017. ISSN 1932–8494. Submitted. [DOI] [PubMed] [Google Scholar]
- Giorgi D, Biasotti S, and Paraboschi L Shape retrieval contest 2007: Watertight models track, 2007. [Google Scholar]
- Gower JC and Ross GJS Minimum spanning trees and single linkage cluster analysis (ref: 69v18 p106–110). Applied Statistics, 18:54–64, 1969. [Google Scholar]
- Huang Q and Guibas LJ Consistent shape maps via semidefinite programming. Comput. Graph. Forum, 32 (5):177–186, 2013. [Google Scholar]
- Huang Q, Koltun V, and Guibas L Joint shape segmentation with linear programming. ACM Trans. Graph, 30 (6):125:1–125:12, December 2011. [Google Scholar]
- Huang Q, Zhang G, Gao L, Hu S, Butscher A, and Guibas LJ An optimization approach for extracting and encoding consistent maps in a shape collection. ACM Trans. Graph, 31(6):167:1–167:11, 2012. [Google Scholar]
- Huang Q, Wang F, and Guibas LJ Functional map networks for analyzing and exploring large shape collections. ACM Trans. Graph, 33(4):36:1–36:11, 2014. [Google Scholar]
- Huang X, Liang Z, Bajaj C, and Huang Q Translation synchronization via truncated least squares. In NIPS, 2017. [PMC free article] [PubMed] [Google Scholar]
- Huber DF and Hebert M Fully automatic registration of multiple 3d data sets. Image and Vision Computing, 21: 637–650, 2001. [Google Scholar]
- Kim V, Li W, Mitra N, DiVerdi S, and Funkhouser T Exploring collections of 3d models using fuzzy correspondences. ACM Trans. Graph, 31(4):54:1–54:11, July 2012. [Google Scholar]
- Kim VG, Lipman Y, and Funkhouser T Blended intrinsic maps. ACM Trans. Graph, 30(4):79:1–79:12, July 2011. [Google Scholar]
- Kraevoy V and Sheffer A Cross-parameterization and compatible remeshing of 3d models. In ACM SIGGRAPH 2004 Papers, SIGGRAPH; ‘04, pp. 861–869, 2004. [Google Scholar]
- Le CM, Levina E, and Vershynin R Concentration and regularization of random graphs. Random Struct. Algorithms, 51(3):538–561, 2017. [Google Scholar]
- Lederman RR and Singer A A representation theory perspective on simultaneous alignment and classification. arXiv preprint arXiv:1607.03464, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lei J and Rinaldo A Consistency of spectral clustering in stochastic block models. Annals of Statistics, 43(1): 215–237, February 2015. [Google Scholar]
- Leonardos S, Zhou X, and Daniilidis K Distributed consistent data association via permutation synchronization. In ICRA, pp. 2645–2652. IEEE, 2017. [Google Scholar]
- Liu C, Yuen J, and Torralba A Sift flow: Dense correspondence across scenes and its applications. IEEE Trans. Pattern Anal. Mach. Intell, 33(5):978–994, 2011. [DOI] [PubMed] [Google Scholar]
- Lowe DG Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision, 60(2):91–110, nov 2004. [Google Scholar]
- Luxburg U A tutorial on spectral clustering. Statistics and Computing, 17(4):395–416, December 2007. [Google Scholar]
- Mandad M, Cohen-Steiner D, Kobbelt L, Alliez P, and Desbrun M Variance-minimizing transport plans for inter-surface mapping. ACM Transactions on Graphics, 36(4), 2017. [Google Scholar]
- Ng AY, Jordan MI, and Weiss Y On spectral clustering: Analysis and an algorithm. In Advances in neural information processing systems, pp. 849–856, 2002. [Google Scholar]
- Nguyen A, Ben-Chen M, Welnicka K, Ye Y, and Guibas LJ An optimization approach to improving collections of shape maps. Comput. Graph. Forum, 30 (5):1481–1491, 2011. [Google Scholar]
- Oliva A and Torralba A Modeling the shape of the scene: A holistic representation of the spatial envelope. Int. J. Comput. Vision, 42(3):145–175, May 2001. [Google Scholar]
- O’Rourke S, Vu V, and Wang K Eigenvectors of random matrices: A survey. J. Comb. Theory, Ser. A, 144:361–442, 2016. [Google Scholar]
- Ovsjanikov M, Ben-Chen M, Solomon J, Butscher A, and Guibas LJ Functional maps: a flexible representation of maps between shapes. ACM Trans. Graph, 31(4): 30:1–30:11, 2012. [Google Scholar]
- Pachauri D, Kondor R, and Singh V Solving the multi-way matching problem by permutation synchronization. In NIPS, pp. 1860–1868, 2013. [Google Scholar]
- Pauly M, Mitra NJ, Giesen J, Gross M, and Guibas LJ Example-based 3d scan completion. In Proceedings of the Third Eurographics Symposium on Geometry Processing, SGP; ‘05, 2005. [Google Scholar]
- Rohe K, Chatterjee S, and Yu B Spectral clustering and the high-dimensional stochastic blockmodel. The Annals of Statistics, 39(4):1878–1915, August 2011. [Google Scholar]
- Rustamov RM Laplace-beltrami eigenfunctions for deformation invariant shape representation. In Proceedings of the Fifth Eurographics Symposium on Geometry Processing, SGP; ‘07, pp. 225–233, 2007. [Google Scholar]
- Schreiner J, Asirvatham A, Praun E, and Hoppe H Inter-surface mapping. In ACM SIGGRAPH 2004 Papers, SIGGRAPH; ‘04, pp. 870–877, 2004. [Google Scholar]
- Shen Y, Huang Q, Srebro N, and Sanghavi S Normalized spectral map synchronization. In NIPS, pp. 4925–4933, 2016. [Google Scholar]
- Snavely N, Seitz SM, and Szeliski R Photo tourism: Exploring photo collections in 3d. ACM Trans. Graph, 25(3):835–846, July 2006. [Google Scholar]
- van Kaick O, Zhang HR, Hamarneh G, and Cohen-Or D A survey on shape correspondence. In EuroGraphics: State-of-the-Art Report, pp. 61–82, 2010. [Google Scholar]
- Wang F, Huang Q, and Guibas LJ Image co-segmentation via consistent functional maps. In ICCV, pp. 849–856, 2013. [Google Scholar]
- Wang F, Huang Q, Ovsjanikov M, and Guibas LJ Unsupervised multi-class joint image segmentation. In CVPR, pp. 3142–3149. IEEE Computer Society, 2014. [Google Scholar]
- Wang L and Singer A Exact and stable recovery of rotations for robust synchronization. Information and Inference: A Journal of the IMA, 2:145193, December 2013. [Google Scholar]
- Zhou T, Lee YJ, Yu SX, and Efros AA Flowweb: Joint image set alignment by weaving consistent, pixel-wise correspondences. In CVPR, pp. 1191–1200. IEEE Computer Society, 2015a. [Google Scholar]
- Zhou X, Zhu M, and Daniilidis K Multi-image matching via fast alternating minimization. CoRR, abs/1505.04845, 2015b. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.