Abstract
One of the main objectives in the analysis of a high dimensional large data set is to learn its geometric and topological structure. Even though the data itself is parameterized as a point cloud in a high dimensional ambient space ℝp, the correlation between parameters often suggests the “manifold assumption” that the data points are distributed on (or near) a low dimensional Riemannian manifold ℳd embedded in ℝp, with d ≪ p. We introduce an algorithm that determines the orientability of the intrinsic manifold given a sufficiently large number of sampled data points. If the manifold is orientable, then our algorithm also provides an alternative procedure for computing the eigenfunctions of the Laplacian that are important in the diffusion map framework for reducing the dimensionality of the data. If the manifold is non-orientable, then we provide a modified diffusion mapping of its orientable double covering.
Keywords: Diffusion maps, Orientability, Dimensionality reduction
1. Introduction
The analysis of massive data sets is an active area of research with applications in many diverse fields. Compared with traditional data analysis, we now analyze larger data sets with more parameters; in many cases, the data is also noisy. Dealing with high dimensional, large, and possibly noisy data sets is the main focus of this area. Among many approaches, dimensionality reduction plays a central role. Its related researches and applications can be seen in areas such as statistics, machine learning, sampling theory, image processing, computer vision and more. Based on the observation that many parameters are correlated to each other, from the right scale of observation, a popular assumption is that the collected data is sampled from a low dimensional manifold embedded in the high dimensional ambient Euclidean space. Under this assumption, many algorithms have been proposed to analyze the data. Some examples are local linear embedding [14], Laplacian eigenmaps [2], Hessian eigenmaps [8], ISOMAP [17], and Diffusion Maps [6]. These nonlinear methods are often found to be superior to traditional linear dimensionality reduction methods, such as principal component analysis (PCA) and classical multidimensional scaling (MDS), since they are able to capture the global nonlinear structure while preserving the local linear structures. Besides dimensionality reduction, data analysts are often faced with problems such as classification, clustering and statistical inference. Clearly, the characteristics of the manifold are useful to solve such problems. A popular approach to extract the spectral geometry of the manifold is by constructing the Laplace operator and examining its eigenfunctions and eigenvalues [6]. The usefulness of this approach is validated by a series of theoretical works [12, 3, 9, 6, 16].
In this paper we consider the following question: Given n data points x1, x2, …, xn sampled from a low-dimensional manifold ℳd but viewed as points in a high dimensional ambient space ℝp, is it possible to determine whether or not the manifold is orientable? For example, can we decide that the Möbius band is non-orientable from just observing a finite number of sampled points? We introduce an algorithm that successfully determines if the manifold is orientable given a large enough number n of sampling points. Our algorithm is shown to be robust to noise in the sense that the data points are allowed to be located slightly off the manifold. If the manifold is determined to be orientable, then a byproduct of the algorithm are eigenvectors that can be used for dimensionality reduction, in an analogous way to the diffusion map framework. The algorithm, referred to as Orientable Diffusion Map (ODM), is summarized in Algorithm 1. If the manifold is determined to be non-orientable, then we show how to obtain a modified diffusion map embedding of its orientable double covering using the odd eigenfunctions of the Laplacian over the double cover.
2. The ODM Algorithm
In this section we give a full description of ODM. There are three main ingredients to our algorithm: local PCA, alignment and synchronization.
2.1. Local PCA
The first ingredient of the ODM algorithm is local PCA. For every data point xi we search for its N nearest neighbors xi1, xi2, …, xiN, where N ≪ n. This set of neighbors is denoted 𝒩xi. The nearest neighbors of xi are located near the tangent space Txiℳ to the manifold at xi, where deviations are possible due to curvature and noise. The result of PCA is an orthonormal basis to a d-dimensional subspace of ℝp which approximates the basis of Txiℳ, provided that N ≥ d + 1. The basis found by PCA is represented as a p × d matrix Oi whose columns are orthonormal, i.e, .
One problem that we may face in practice is lack of knowledge of the intrinsic dimension d. Estimating the dimensionality of the manifold from sampled points is an active area of research nowadays, and a multiscale version of the local PCA algorithm is presented and analyzed in [11]. We remark that such methods for dimensionality estimation can be easily incorporated in the first step of our ODM algorithm. Thus, we may allow N to vary from one data point to another. In this paper, however, we use the classical approach to local PCA, since it simplifies the presentation. We remark that there is no novelty in our usage of local PCA, and our detailed exposition of this step is merely for the sake of making this paper as self contained as possible. Readers who are familiar with PCA may quickly move on to the next subsection.
Algorithm 1.
Require: Data points x1, x2, …, xn ∈ ℝp.
|
In our approach, we try to estimate the intrinsic dimension from the singular values of the mean-shifted local data matrix
of size p × N, where . We denote the singular values of Xi by σi,1 ≥ σi,2 ≥ … ≥ σi,N. If the neighboring points in 𝒩xi are located exactly on Txiℳ, then rank Xi = d, and there are only d non-vanishing singular values (i.e., σi,d+1 = … = σi,N = 0). In such a case, the dimension can be estimated as the number of non-zero singular values. In practice, however, due to the curvature effect, there may be more than d non-zero singular values. A common practice is to estimate the dimension as the number of singular values that account for high enough percentage of the variability of the data. That is, one sets a threshold γ between 0 and 1 (usually closer to 1 than to 0), and estimates the dimension as the smallest integer di for which
For example, setting γ = 0.9 means that di singular values account for at least 90% variability of the data, while di − 1 singular values account for less than 90%. We refer to the smallest integer di as the estimated local dimension of ℳ at xi. One possible way to estimate the dimension of the manifold would be to use the mean of the estimated local dimensions d1, …, dn, that is, (and then round it to the closest integer). The mean estimator minimizes the sum of squared errors . We estimate the intrinsic dimension of the manifold by the median value of all the di’s, that is, we define the estimator d̂ for the intrinsic dimension d as
The median has the property that it minimizes the sum of absolute errors (originally due to Laplace). As a result, estimating the intrinsic dimension by the median is more robust to outliers compared to the mean estimator. In all proceeding steps of the algorithm we use the median estimator d̂, but in order to facilitate the notation we write d instead of d̂.
Suppose that the SVD of Xi is given by
The columns of the p × N matrix Ui are orthonormal and are known as the left singular vectors
We define the matrix Oi by the first d left singular vectors (corresponding to the largest singular values):
(1) |
Note that the columns of Oi form a numerical approximation to an orthonormal basis of the tangent plane Txiℳ.
2.2. Alignment
The second ingredient of the ODM algorithm is alignment. Suppose xi and xj are two nearby points, satisfying1 xj ∈ 𝒩xi and xi ∈ 𝒩xj. Since N ≪ n, the curvature effect is small, and the tangent spaces Txiℳ and Txjℳ are also close2. As a result, the matrix is not necessarily orthogonal, and we define Oij as its closest orthogonal matrix, i.e.,
(2) |
where ‖ · ‖F is the Hilbert-Schmidt norm (also known as the Frobenius norm). The minimization problem (2) has a simple solution via the singular value decomposition (SVD) of [1]. Specifically, if
(3) |
is the SVD of , then
(4) |
solves (2). The optimal orthogonal transformation Oij can be either a rotation (i.e., an element of SO(d)), or a composition of a rotation and a reflection. Figure 1 is an illustration of this procedure.
The determinant of Oij classifies the two possible cases:
(5) |
We refer to the process of finding the optimal orthogonal transformation between bases as alignment. Note that not all bases are aligned; only the bases of nearby points are aligned. We set G = (V, E) to be the undirected graph with n vertices corresponding to the data points, where an edge between i and j exists iff their corresponding bases are aligned by the algorithm. We further encode the information about the reflections in a symmetric n × n matrix Z whose elements are given by
(6) |
That is, Zij = 1 if no reflection is needed, Zij = −1 if a reflection is needed, and Zij = 0 if the points are not nearby.
2.3. Synchronization
The third ingredient of the ODM algorithm is synchronization. If the manifold is orientable, then we can assign an orientation to each tangent space in a continuous way. For each xi, the basis found by local PCA can either agree or disagree with that orientation. We may therefore assign a variable zi ∈ {+1, −1} that designates if the PCA basis at xi agrees with the orientation (zi = 1) or not (zi = −1). Clearly, for nearby points xi and xj whose bases are aligned, we must have that
(7) |
That is, if no reflection was needed in the alignment stage (i.e., zij = 1), then either both PCA bases at xi and xj agree with the orientation (zi = zj = 1) or both disagree with the orientation (zi = zj = −1). Similarly, if a reflection was needed (zij = −1), then it must be the case that zi = −zj. In practice, however, we are not given an orientation for the manifold, so the values of the zi’s are unknown. Instead, only the values of the zij’s are known to us after the alignment stage. Thus, we may try solving the system (7) to find the zi’s. A solution exists iff the manifold is orientable. The solution is unique up to a global sign change, i.e., if (z1, z2, …, zn) is a solution, then (−z1, −z2, …, −zn) is also a solution. If all equations in (7) are accurate and the graph G is connected, then finding a solution can be obtained in a straightforward manner by simply traversing a spanning tree of the graph from the root to the leafs.
In practice, however, due to curvature effects and noise, some of the equations in (7) may be incorrect and we would like to find a solution that satisfies as many equations as possible. This maximum satisfiability problem is also known as the synchronization problem over the group ℤ2 [15], and in [15, 7] we proposed spectral algorithms to approximate its solution. In [15] we used the top eigenvector (corresponding to the largest eigenvalue) of Z to find the solution, while in [7] we normalized the matrix Z and used the top eigenvector of that normalized matrix. Here we use the algorithm in [7], since we find the normalization to give much improved results.
Specifically, we normalize the matrix Z by dividing the elements of each row by the number of non-zero elements in that given row, that is,
(8) |
where D is a n × n diagonal matrix with . In other words, Dii = deg(i), where deg(i) is the degree of vertex i in the graph G. We refer to 𝒵 as the reflection matrix. Note that although 𝒵 is not necessarily symmetric, it is similar to the symmetric matrix D−1/2Z D−1/2 through
Therefore, the matrix 𝒵 has n real eigenvalues and n orthonormal eigenvectors , satisfying
We compute the top eigenvector of 𝒵, which satisfies
(9) |
and use it to obtain estimators ẑ1, …, ẑn for the unknown variables z1, …, zn, in the following way:
(10) |
The estimation by the top eigenvector is up to a global sign, since if is the top eigenvector of 𝒵 then so is .
To see why (10) is a good estimator, we first analyze it under the assumption that the matrix Z contains no errors on the relative reflections. Denoting by ϒ the n × n diagonal matrix with ±1 on its diagonal representing the correct reflections zi, i.e. ϒii = zi, we can write the matrix Z as
(11) |
where A is the adjacency matrix of the graph G whose elements are given by
(12) |
because in the error-free case for (i, j) ∈ E. The reflection matrix 𝒵 can now be written as
(13) |
Hence, 𝒵 and D−1 A share the same eigenvalues. Since the normalized discrete graph Laplacian ℒ of the graph G is defined as
(14) |
it follows that in the error-free case, the eigenvalues of I − 𝒵 are the same as the eigenvalues of ℒ. These eigenvalues are all non-negative, since ℒ is similar to the positive semidefinite matrix I − D−1/2 AD−1/2, whose non-negativity follows from the identity
for any υ ∈ ℝn. In other words,
(15) |
where the eigenvalues of ℒ are ordered in increasing order, i.e., , and the corresponding eigenvectors satisfy . Furthermore, the sets of eigenvectors are related by
(16) |
If the graph G is connected, then the eigenvalue is simple and its corresponding eigenvector is the all-ones vector 1 = (1, 1, …, 1)T. Therefore,
(17) |
and, in particular,
(18) |
This implies that in the error-free case, (10) perfectly recovers the reflections. In other words, the eigenvector contains the orientation information of each tangent space. According to this orientation information, we can synchronize the orientations of all tangent spaces by flipping the orientation of the basis Oi whenever . In conclusion, the above steps synchronize all the tangent spaces to agree on their orientations.
If the manifold is not orientable, then it is impossible to synchronize the orientation of the tangent spaces and the top eigenvector cannot be interpreted as before. We return to this interpretation later in the paper. For the moment, we just remark that the eigenvalues of A and Z (and similarly of ℒ and 𝒵) contain information about the number of orientation preserving and orientation reversing closed loops. To that end, consider first At(i, j), which is the number of paths of length t over the graph G that start at node i and end at node j. In particular, At(i, i) is the number of closed loops that start and end at node i. Therefore,
is the number of closed loops of length t. Similarly,
is the number of orientation preserving closed loops minus the number of orientation reversing closed loops of length t. If the manifold is orientable then all closed loops are orientation preserving and the two traces are the same, but if the manifold is non-orientable then there are orientation reversing closed loops and therefore the eigenvalues of A and Z must differ.
If the manifold is determined to be orientable, then we can use the computed eigenvectors of 𝒵 for dimensionality reduction in the following manner. Specifically, we use point-wise multiplication of the eigenvectors of 𝒵 by the estimated reflections to produce new vectors υ̃1, …, υ̃n that are given by
(19) |
The relation (16) between the eigenvectors of 𝒵 and ℒ, and the particular form of the top eigenvector given in (17) imply that the vectors υ̃1, …, υ̃n are eigenvectors of ℒ with . The eigenvectors of ℒ are used in the diffusion map framework to embed the data points onto a lower dimensional Euclidean space, where distances between points approximate the diffusion distance [6]. Similarly, we can embed the data points using the following rule:
(20) |
where t ≥ 0 is a parameter.
If the data points are sampled from the uniform distribution over the manifold, then the eigenvectors of the discrete graph Laplacian ℒ approximate the eigenfunctions of the Laplace-Beltrami operator [4], and it follows that the vectors υ̃i that are used in our ODM embedding (20) also converge to the eigen-functions of the Laplace-Beltrami operator. If the sampling process is from a different distribution, then the υ̃i’s approximate the eigenfunctions of the Fokker-Planck operator. By adapting the normalization rule suggested in [6] to our matrix Z, we can get an approximation of the eigenfunctions of the Laplace-Beltrami operator also for non-uniform distributions.
We remark that nothing prevents us from running the ODM algorithm in cases for which the sampling process from the manifold is noisy. We demonstrate the robustness of our algorithm through several numerical experiments in Section 4.
3. Orientable double covering
Every non-orientable manifold has an orientable double cover. Can we reconstruct the double covering from data points sampled only from the non-orientable manifold? In this section we give an affirmative answer to this question, by constructing a modified diffusion map embedding of the orientable double cover using only points that are sampled from the non-orientable manifold.
For simplicity of exposition, we start by making a simplifying assumption that we shall later relax. The assumption is that the orientable double cover has a symmetric isometric embedding in Euclidean space. That is, we assume that if ℳ is the non-orientable manifold, and ℳ̃ is its orientable double cover, then ℳ̃ has a symmetric isometric embedding in ℝq, for some q ≥ 1. Symmetric embedding means that for all x ∈ ℳ̃ ⊂ ℝq, also −x ∈ ℳ̃. This assumption is motivated by examples: the orientable double cover of the Möbius band is the cylinder, the orientable double cover of ℝP2 is S2 and the orientable double cover of the Klein bottle is T2. The cylinder, S2 and T2 all have symmetric isometric embeddings in ℝ3 (q = 3). We believe that this assumption holds true in general, but its proof is beyond the scope of this paper3. Our assumption means that we have a ℤ2 action on ℳ̃ given by x ↦ −x. We note that this ℤ2 action is isometric and recall that if the group ℤ2 is properly discontinuously acting on ℳ̃, then ℳ̃/ℤ2 is non-orientable if and only if ℳ̃ has an orientation that is not preserved by the ℤ2 action.
By our assumption, the orientable double covering ℳ̃ can be isometrically embedded into ℝq for some q so that ℤ2x = −x for all x ∈ ℳ̃ and ℳ̃/ℤ2 = ℳ. Denote the projection map as π : {x, −x} ⊂ ℳ̃ ↦ [x] ∈ ℳ. Under this assumption, each pair of antipodal points in ℳ̃ can be viewed as a single point in ℳ. In other words, for any point [x] ∈ ℳ, we can view it as two antipodal points x, −x ∈ ℳ̃ whose antipodal relationship is ignored.
We can now offer an interpretation of the matrix Z whose entries are given in (6). Recall that we have n data points in ℝp that are sampled from ℳ. We denote these points by [x1], …, [xn]. Associate to each [xi] ∈ ℳ a point xi ∈ ℳ̃ ⊂ ℝq such that π(xi) = [xi] and such that the entries of Z satisfy the following conditions: zij = 1 if xj is in the ϵ-neighborhood of xi; zij = −1 if xj is the ϵ-neighborhood of ℤ2xi = −xi; and zij = 0 if the distance between xj and xi and the distance between xj and −xi are both larger than ϵ. That is,
Next, we build a 2n × 2n matrix Z̃ as follows:
(21) |
where ⊗ is the Kronecker product. The matrix Z̃ corresponds to 2n points over ℳ̃, while originally we had only n such points. We identify the additional points, denoted xn+1, …, x2n as xn+i = −xi for i = 1, …, n. With this identification, it is easily verified that the entries of are also given by
In a sense, the −Z in the (1, 2) and (2, 1) entries of Z̃ in (21) provides the lost information between π−1([xi]) and π−1([xj]). We also define the matrices 𝒵̃ and ℒ̃ as
(22) |
and
(23) |
It is clear that the matrix 𝒵̃ is a discretization4 of the integral operator 𝒦̃ϵ : L2(ℳ̃) → L2(ℳ̃), given by
(24) |
where the kernel function K̃ϵ is given by
(25) |
Bϵ(x) is a ball of radius ϵ in ℝq centered at x, and the volume is measured using the metric of ℳ̃ induced from ℝq. Clearly, the integral operator 𝒦̃ϵ commutes with the ℤ2 action:
(26) |
where (ℤ2f)(x) = f(−x). Moreover, the kernel of 𝒦̃ϵ includes all “even” functions. That is, if f satisfies f(x) = f(−x) for all x ∈ ℳ̃, then 𝒦̃ϵ f = 0. It follows that the eigenfunctions of 𝒦̃ϵ with non-vanishing eigenvalues must be “odd” functions that satisfy f(x) = −f(−x).
Recall that the graph Laplacian ℒ converges to the Laplace-Beltrami operator in the limit n → ∞ and ϵ → 0 if the data points are sampled from the uniform distribution over the manifold. That is,
(27) |
where m2 is a constant related to the second moment of the kernel function. Adapting this result to our case implies that we have the following convergence result for the matrix ℒ̃:
(28) |
where m̃2 is a constant. In other words, in the limit we recover the following operator:
(29) |
The Laplacian Δℳ̃ also commutes with the ℤ2 action, so by similar considerations, the kernel of Δℳ̃ − ℤ2 Δℳ̃ contains all even functions, and the eigen-functions with non-zero eigenvalues are odd functions.
Suppose ϕ1, ϕ2, …, ϕ2n and λ1 ≥ λ2 ≥ … ≥ λ2n ≥ 0 are the eigenvectors and eigenvalues of ℒ̃. From the above discussion we have that λn+1 = … = λ2n = 0 and ϕ1, …, ϕn are odd vectors satisfying ϕl(i) = −ϕl(n+i) (1 ≤ l ≤ n). We define the map Φ̃t for t > 0 as
(30) |
We refer to Φ̃t as a modified diffusion map. Note that Φ̃t uses only the odd eigenfunctions of the Laplacian over ℳ̃, whereas the classical diffusion map uses all eigenfunctions, both odd and even.
Our construction shows that from the information contained in the matrix Z, which is built solely from the non-orientable manifold ℳ, we obtain the odd eigenfunctions of the Laplacian over the double cover ℳ̃. The odd eigenfunctions are extremely useful for embedding the double cover. For example, the double cover of ℝP2 is S2 whose odd eigenfunctions are the odd degree spherical harmonics. In particular, the degree-1 spherical harmonics are the coordinate functions x, y and z (in ℝ3), so truncating the modified diffusion map Φ̃t to ℝ3 results in an embedding of S2 in ℝ3. Similarly, the first 3 eigenvectors of ℒ̃ (or equivalently, 𝒵̃) built up from the Klein bottle and the Möbius band allow us to get a torus and a cylinder, respectively. These reconstruction results are shown in Section 4.
We now return to the underlying assumption that the manifold ℳ̃ has a symmetric isometric embedding in ℝq for some q ≥ 1. While the assumption was useful to visualize the construction, we note that it is actually not necessary for the above derivation and interpretation. Indeed, the non-orientable manifold ℳ has an abstract orientable double cover ℳ̃ with ℤ2 acting on it such that ℳ̃/ℤ2 = ℳ. While the assumption allowed us to make a concrete specification of the ℤ2 action (i.e., ℤ2x = −x), this specification is not necessary in the abstract setup. The careful reader may check each and every step of the derivation and verify that they follow through also in the abstract sense. In particular, the matrix 𝒵̃ can be interpreted just as before, with xn+i = ℤ2xi ∈ ℳ̃ for ℤ2 being the abstract action. Moreover, the Laplacian over ℳ̃ still commutes with ℤ2 and the matrix ℒ̃ converges to the operator given in (29). Thus, the modified diffusion mapping Φ̃t given in (30) is an embedding of the abstract double cover. In the continuous setup, the eigenfunctions of the operator Δℳ̃ − ℤ2Δℳ̃ embed the abstract double cover ℳ̃ in the Hilbert space 𝓁2 through the mapping
(31) |
where
Finally, we show a useful “trick” to generate a point cloud over a non-orientable manifold without knowing any of the parametrizations that embed it in Euclidean space. For example, suppose we want to generate a point cloud over the Klein bottle, but we forgot its parametrization in ℝ4 (the parametrization is given later in (32)). There is still a rather simple way to proceed that only requires the knowledge of the orientable double cover. That is, given a non-orientable manifold ℳ from which we want to sample points in a Euclidean space, all we need to know is its orientable double cover ℳ̃ and its symmetric embedding in Euclidean space5. So, although we forgot the parametrization of the Klein bottle in ℝ4, we assume that we still remember that its double cover is T2 and the parametrization of T2 in ℝ3. From the symmetric embedding of T2 in ℝ3, we now describe how to get a diffusion map embedding of the Klein bottle in Euclidean space. In general, from the symmetric embedding of ℳ̃ we describe how to get the diffusion map embedding of ℳ.
We start by sampling n points uniformly from ℳ̃ ∈ ℝq. Next, we double the number of points by setting xn+i = −xi for i = 1, …, n. We then build a 2n × 2n matrix given by
We see that the matrix à does not distinguish between x and ℤ2x, which leads to the previous identification {x, ℤ2x} = [x] ∈ ℳ. After normalizing the matrix as before, it can be easily verified that it converges to an integral operator that has all the odd functions in its kernel and its non-zero eigenvalues correspond to even eigenfunctions. Moreover, in the limit of n → ∞ and ϵ → 0 we get convergence to the following operator:
This means that the computed eigenvectors approximate the even eigenfunctions of the Laplacian over ℳ̃. That is, we compute eigenfunctions that satisfy ϕl(x) = ϕl(ℤ2x) (for l = 1, …, n). As a result, these eigenfunctions are only functions of [x] and are the eigenfunctions of Δℳ. Thus, the mapping
is the diffusion mapping of ℳ.
We remark that this construction can help us to build up more examples of non-orientable manifolds. For example, we can start from S2d ⊂ ℝ2d+1 and build up the diffusion map embedding of the non-orientable real projective space ℝP(2d) without knowing the parametrization of ℝP(2d).
4. Numerical examples
In this section we provide several numerical experiments demonstrating the application of ODM on simulative data. In our experiments, we measure the level of noise by the signal to noise ratio (SNR), which is defined in the following way. Suppose s1, s2, …, sn ∈ ℝp is a set of “clean” data points that are located exactly on the manifold. This set of points is considered as the signal, and we define its sample variance as
where is the sample mean. The set of noisy data points x1, x2, …, xn are located off the manifold and are generated using the following process:
where σ > 0 is a fixed parameter, and ξ ~ 𝒩 (0, Ip×p) are independently identically distributed (i.i.d.) standard multivariate normally distributed isotropic random variables. The SNR is measured in dB and is defined as
Our first example is the unit sphere S2. This example demonstrates how to synchronize the orientation when the manifold is orientable. We sample n = 5000 points from S2 uniformly without noise and run ODM with N = 100 nearest neighbors, and estimated the intrinsic dimension in the local PCA step using γ = 0.8. Then we sample another 5000 points from S2 with SNR=5dB and run ODM with the same parameters. In both cases, the intrinsic dimension is estimated correctly as d̂ = 2. The histograms of the top eigenvector of both cases are shown in Figure 2. The results show the robustness of ODM in orientation synchronization.
Next we show how the ODM algorithm helps us to determine if a given manifold is orientable or not. We sample 5000 points from the Klein bottle embedded in ℝ4 and ℝP(2) embedded in ℝ4 without noise, respectively. Then run the ODM algorithm for these two data sets with N = 200 and γ = 0.8. The histograms of the top eigenvectors are shown in Figure 3. It shows that when the underlying manifold is non-orientable, the values of the first eigenvector of the ODM do not cluster around two distinguished positive and negative values.
We proceed to considering the application of ODM for dimensionality reduction, looking at data points sampled from the bell-shaped, closed, non-intersected curve, which is parameterized by C(θ) = (x(θ), y(θ)), θ ∈ [0, 2π], where
The curve is depicted in Figures 4 and 5. Note that the isthmus of C(θ) imposes a difficultly whenever the noise is too large or the sampling rate is too low. This difficulty is due to the identity of the nearest neighbors, as illustrated in Figure 4. It reflects the fact that the nearest neighbors are neighbors in the extrinsic sense (when the distance is measured by the Euclidean distance), but not in the intrinsic sense (when the distance is measured by the geodesic distance). To cope with this difficulty, we first cluster the nearest neighbors in order to find the “true” neighbors in the geodesic distance sense, prior to applying PCA. For clustering the points in 𝒩xi we apply the spectral clustering algorithm [13], and for PCA we consider only the points that belong to the cluster of xi.
We applied ODM to 5000 points from C(θ) without noise, with parameters N = 200 and γ = 0.8, and then embed the data points into ℝ2 using the second and third eigenvectors and (20). We repeat the same steps for noisy data sets with SNR=30dB and SNR=24dB. Figure 5 shows the original data set, the diffusion map (DM) embedding, and the ODM embedding. In the DM algorithm, we used the Gaussian kernel exp{−‖xi − xj‖2/σ} with σ = 0.5. As expected, the embedding results of ODM and DM are comparable.
Our next example is the Swiss roll, a manifold often used by machine learners as a testbed for dimensionality reduction methods. The Swiss roll model is the following manifold S embedded in ℝ3:
for (t, s) ∈ [0, 1] × [0, 21]. The Swiss roll model exhibits the same difficulty we encountered before with the bell-shape curve, namely, that the geodesically distant points might be close in the Euclidean sense (see Figure 6). Moreover, unlike the manifolds considered before, the Swiss roll has a boundary. To cope with these problems, we again apply the spectral clustering algorithm [13] prior to the local PCA step of the ODM algorithm.
We sampled n = 7000 points from the Swiss roll without noise, run ODM with parameters N = 40 and γ = 0.8, and then embedded the data points into ℝ2 by the second and third eigenvectors using (20). We repeated the same steps for data sets with SNR values 50dB, 30dB and 25dB. Figure 7 shows the original data, the embedding by DM, and the embedding by ODM. In the DM algorithm, we used the Gaussian kernel with σ = 5.
In conclusion, we see that besides determining the orientability of the manifold, ODM can also be used as a dimensionality reduction method.
In the last part of this section we demonstrate results about recovering the orientable double covering of a non-orientable manifold as discussed in Section 3. First we demonstrate the orientable double covering of ℝP(2), Klein bottle and Möbius band in Figure 8. For ℝP(2), we use the following embedding inside ℝ4:
where (x, y, z) ∈ S2 ⊂ ℝ3. We sample 8000 points uniformly from S2 and map the sampling points to ℝP(2) by the embedding, which gives us a non-uniform sampling data set. Then we evaluate eigenvectors of the 16000 × 16000 sparse matrix 𝒵̃ constructed from the data set with N = 40 and get the approximation of the coordinates of the orientable double covering which allows us to get the orientable double covering (Figure 8, left panel).
For the Klein bottle, we apply the following embedding inside ℝ4:
(32) |
where (u, υ) ∈ [0, 2π] × [0, 2π]. We sample 8000 points uniformly from [0, 2π] × [0, 2π] and map the sampling points to the Klein bottle by the embedding, which gives us a non-uniform sampling data set over the Klein bottle. Then we run the same procedure as in ℝP(2) with N = 40 and get the result (Figure 8, middle panel).
For the Möbius band, we use the following parametrization:
where a = 3, b = 5 and (u, υ) ∈ [−1, 1]×[0, 2π]. As above, we sample 8000 points uniformly from [−1, 1]×[0, 2π] and map the sampling points to the Möbius band by the parametrization, which gives us a non-uniform sampling data set. Then we run the same procedure with N = 20 and get the result (Figure 8, right panel). It is worth mentioning that when b is small, e.g., b = 1, we fail to recover the cylinder, as can be seen in Figure 9. This difficulty comes from the lack of information in the width of the band when b is too small.
Last, we demonstrate how to get a non-orientable manifold by constructing the 𝒜̃ matrix as discussed in Section 3. We start from showing the spectrum of ℝP(2) constructed by gluing the antipodal points of S2 together. We build up 𝒜̃ by taking N = 50 from 16000 data points sampled from S2 embedded in ℝ3:
which is the approximation of the Laplace-Beltrami operator over S2/ℤ2 ≅ ℝP(2). The spectrum of the operator is shown in Figure 10.
Then we demonstrate the reconstruction of the orientable double covering of an abstract non-orientable manifold, by which we mean an abstract non-orientable manifold not yet embedded inside any Euclidean space. We start from sampling 16000 points from a cylinder C embedded in ℝ3, which is parameterized by:
where (θ, a) ∈ [0, 2π]×[−2, 2], and run the following steps. First we fix N = 50, and build up 𝒜̃, the approximation of the Laplace-Beltrami operator over C/ℤ2 by gluing the antipodal points together. Note that C/ℤ2 is an abstract Möbius band up to now. Second, we embed C/ℤ2 by applying diffusion map, where we take the second to fifth eigenvectors of 𝒜̃, that is, we embed C/ℤ2 into ℝ4. Finally we run ODM with N = 50 over the embedded C/ℤ2 to reconstruct the double covering. The result, shown in Figure 11, shows that we recover back a cylinder. Similarly, we run the same process with the same parameters N = 50 and the second to fifth eigenvectors of 𝒜̃ to embed T2/ℤ2 and S2/ℤ2 into ℝ4, and then run ODM with N = 50, which results in the reconstruction of the orientable double covering of an abstract Klein bottle and ℝP(2). The results are shown in Figure 12 and 13. Using different number of eigenvectors for embedding, for example, from the second to sixth or from the second to seventh eigenvectors give us similar results.
Note that in the above three examples shown in Figures 11, 12 and 13, the recovered objects look different from the original one. This distortion mainly comes from the following step: the embedding by diffusion map chosen in the demonstration is not isometric but just an approximation of an isometric embedding.
5. Summary and Discussion
We introduced an algorithm for determining the orientability of a manifold from scattered data. If the manifold is orientable then the algorithm synchronizes the orientations and can further be used for dimensionality reduction in a similar fashion to the diffusion map framework.
The algorithm consists of three steps: local PCA, alignment and synchronization. In the synchronization step we construct a matrix 𝒵, which for an orientable manifold is shown to be similar to the normalized Laplacian ℒ. For non-orientable manifolds this similarity does not hold anymore and we show how to obtain an embedding of the orientable double covering using the eigenvectors of the Kronecker product . In particular, we show that the eigenvectors of 𝒵̃ approximate the odd eigenfunctions of the Laplacian over the double covering. We also show how to using this observation in order to embed an abstract non-orientable manifold in Euclidean space given its orientable double covering.
We comment that the reflection matrix 𝒵 can also be built by taking a kernel function into consideration, like the weighted graph Laplacians discussed in [6].
For example, we may replace the definition (6) of Z by
(33) |
and
(34) |
where D is a n × n diagonal matrix with . All the results in this paper also apply to this kernelized reflection matrix.
One possible application of our framework is the geometrical and topological study of high contrast local patches that are extracted from natural images [10]. Computational topology methods were used in [5] to conclude that this data set has the topology of the Klein bottle. This result may perhaps be confirmed using the methods presented here.
Finally, we remark that while the reflection matrix only uses the information in det Oij, it seems natural also to take into account all the information in Oij. This modification will be the subject of a separate publication.
Acknowledgements
The problem of reconstructing the double covering of a non-orientable manifold was proposed to us by Peter Jones and Ronald Coifman during a discussion at Yale University. A. Singer was partially supported by Award Number DMS-0914892 from the NSF, by Award Number FA9550-09-1-0551 from AFOSR, by Award Number R01GM090200 from the NIGMS and by the Alfred P. Sloan Foundation. H.-T. Wu acknowledges support by FHWA grant DTFH61-08-C-00028.
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Other notions of proximity are also possible. For example, xi and xj are nearby if they have at least c common points, that is, if |𝒩xi ∩ 𝒩xj | ≥ c, where c ≥ 1 is a parameter.
This means that their Grassmannian distance defined via the operator norm is small.
We elude to a possible proof by noting that Nash embedding theorem, that guarantees that every Riemannian manifold can be isometrically embedded in Euclidean space, is based on modified Newton iterations. Thus, a possible roadmap for the proof would be to show that Whitney embedding theorem gives a symmetric embedding as an initialization, and then to show that the iterations in Nash theorem preserve symmetry.
That is, 𝒵̃ converges to 𝒦̃ϵ in the limit n → ∞.
Here we actually need the assumption that the orientable double cover has a symmetric isometric embedding in Euclidean space.
References
- 1.Arun KS, Huang TS, Blostein SD. Least-squares fitting of two 3-D point sets. IEEE Trans. Patt. Anal. Mach. Intell. 1987;9(5):698–700. doi: 10.1109/tpami.1987.4767965. [DOI] [PubMed] [Google Scholar]
- 2.Belkin M, Niyogi P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation. 2003;15(6):1373–1396. [Google Scholar]
- 3.Belkin M, Niyogi P. Towards a theoretical foundation for Laplacian-based manifold methods. Proceedings of the 18th Conference on Learning Theory (COLT); 2005. pp. 486–500. [Google Scholar]
- 4.Belkin M, Niyogi P. Advances in Neural Information Processing Systems (NIPS) MIT Press; 2007. Convergence of Laplacian eigenmaps. [Google Scholar]
- 5.Carlsson G, Ishkhanov T, de Silva V, Zomorodian A. On the local behavior of spaces of natural images. International Journal of Computer Vision. 2008;76(1):1–12. [Google Scholar]
- 6.Coifman RR, Lafon S. Diffusion maps. Applied and Computational Harmonic Analysis. 2006;21(1):5–30. [Google Scholar]
- 7.Cucuringu M, Lipman Y, Singer A. Sensor network localization by eigenvector synchronization over the Euclidean group. doi: 10.1145/2240092.2240093. Submitted for publication. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Donoho DL, Grimes C. Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data. Proceedings of the National Academy of Sciences of the United States of America. 2003;100(10):5591–5596. doi: 10.1073/pnas.1031596100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Hein M, Audibert J, von Luxburg U. From graphs to manifolds - weak and strong pointwise consistency of graph Laplacians. Proceedings of the 18th Conference on Learning Theory (COLT); 2005. pp. 470–485. [Google Scholar]
- 10.Lee AB, Pedersen KS, Mumford D. The non-linear statistics of high-contrast patches in natural images. International Journal of Computer Vision. 2003;54(13):83–103. [Google Scholar]
- 11.Little AV, Lee J, Jung YM, Maggioni M. Estimation of intrinsic dimensionality of samples from noisy low-dimensional manifolds in high dimensions with multiscale SVD. 2009 IEEE/SP 15th Workshop on Statistical Signal Processing; 2009. pp. 85–88. [Google Scholar]
- 12.Nadler B, Lafon S, Coifman RR, Kevrekidis IG. Diffusion maps, spectral clustering and reaction coordinates of dynamical systems. Applied and Computational Harmonic Analysis. 2006;21(1):113–127. [Google Scholar]
- 13.Ng AY, Jordan MI, Weiss Y. On spectral clustering: Analysis and an algorithm. In: Leen TK, Dietterich TG, Tresp V, editors. Advances in neural information processing systems 14. Cambridge, MA: MIT Press; 2002. [Google Scholar]
- 14.Roweis ST, Saul LK. Nonlinear dimensionality reduction by locally linear embedding. Science. 2000;290(5500):2323–2326. doi: 10.1126/science.290.5500.2323. [DOI] [PubMed] [Google Scholar]
- 15.Singer A. Angular synchronization by eigenvectors and semidefinite programming. doi: 10.1016/j.acha.2010.02.001. In press. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Singer A. From graph to manifold Laplacian: The convergence rate. Applied and Computational Harmonic Analysis. 2006;21(1):128–134. [Google Scholar]
- 17.Tenenbaum JB, de Silva V, Langford JC. A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science. 2000;290(5500):2319–2323. doi: 10.1126/science.290.5500.2319. [DOI] [PubMed] [Google Scholar]