Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2024 Jan 25.
Published in final edited form as: Inverse Probl. 2020 Apr 29;36(6):064002. doi: 10.1088/1361-6420/ab7d2c

NON-UNIQUE GAMES OVER COMPACT GROUPS AND ORIENTATION ESTIMATION IN CRYO-EM

Afonso S Bandeira 1, ETH Zürich 2, Yutong Chen 3,*, Roy R Lederman 4, Amit Singer 5
PMCID: PMC10810340  NIHMSID: NIHMS1958378  PMID: 38274355

Abstract

Let 𝒢 be a compact group and let fijC(𝒢). We define the Non-Unique Games (NUG) problem as finding g1,,gn𝒢 to minimize i,j=1nfijgigj-1. We introduce a convex relaxation of the NUG problem to a semidefinite program (SDP) by taking the Fourier transform of fij over 𝒢. The NUG framework can be seen as a generalization of the little Grothendieck problem over the orthogonal group and the Unique Games problem and includes many practically relevant problems, such as the maximum likelihood estimator to registering bandlimited functions over the unit sphere in d-dimensions and orientation estimation of noisy cryo-Electron Microscopy (cryo-EM) projection images. We implement a SDP solver for the NUG cryo-EM problem using the alternating direction method of multipliers (ADMM). Numerical study with synthetic datasets indicate that while our ADMM solver is slower than existing methods, it can estimate the rotations more accurately, especially at low signal-to-noise ratio (SNR).

1991 Mathematics Subject Classification. Primary: 00A69; Secondary: 90C34, 20C40

Key words and phrases. Computer vision, pattern recognition, algorithms, optimization, cryo-EM

1. Introduction

We consider problems of the following form:

minimizeg1,,gni,j=1nfijgigj-1subject togi𝒢, (1.1)

where 𝒢 is a compact group and fij:𝒢R are suitable functions. We will refer to such problems as a Non-Unique Game (NUG) problem over 𝒢.

Note that the solution to the NUG problem is not unique. If g1,,gn is a solution to (1.1), then so is g1g,,gng for any g𝒢. That is, we can solve (1.1) up to a global shift g𝒢.

In many inverse problems, the goal is to estimate multiple group elements from information about group offsets, and can be formulated as (1.1). A simple example is angular synchronization [40], where one is tasked with estimating angles θii from information about their offsets θi-θj mod 2π. The problem of estimating the angles can then be formulated as an optimization problem depending on the offsets, and thus be written in the form of (1.1). In this case, 𝒢SO(2).

One of the simplest instances of (1.1) is the Max-Cut problem, where the objective is to partition the vertices of a graph as to maximize the number of edges (the cut) between the two sets. In this case, 𝒢Z2, the group of two elements {±1}, and fij is zero if (i,j) is not an edge of the graph and

fij(1)=0andfij(-1)=-1,

if (i,j) is an edge. In fact, we take a semidefinite programming based approach towards (1.1) that is inspired by — and can be seen as a generalization of — the semidefinite relaxation for the Max-Cut problem by Goemans and Williamson [21].

Another important source of inspiration is the semidefinite relaxation of Max-2-LinZL, proposed in [15], for the Unique Games problem, a central problem in theoretical computer science [26, 27]. Given integers n and L, an Unique-Games instance is a system of linear equations over ZL on n variables xii=1n. Each equation constraints the difference of two variables. More precisely, for each (i,j) in a subset of the pairs, we associate a constraint

xi-xj=bijmodL.

The objective is then to find xii=1n in ZL that satisfy as many equations as possible. This can be easily described within our framework by taking, for each constraint,

fij(g)=-δgbij,

and fij=0 for pairs not corresponding to constraints. The term ‘unique’ derives from the fact that the constraints have this special structure where the offset can only take one value to satisfy the constraint, and all other values have the same score. This motivated our choice of nomenclature for the framework treated in this paper. The semidefinite relaxation for the unique games problem proposed in [15] was investigated in [8] in the context of the signal alignment problem, where the fij are not forced to have a special structure (but 𝒢ZL). The NUG framework presented in this paper can be seen as a generalization of the approach in [8] to other compact groups 𝒢. We emphasize that, unlike [8] that was limited to the case of a finite cyclic group, here we consider compact groups that are possibly infinite and non-commutative.

Besides the signal alignment problem treated in [8] the semidefinite relaxation to the NUG problem we develop generalizes with other effective relaxations. When 𝒢Z2 it coincides with the semidefinite relaxations for Max-Cut [21], the little Grothendieck problem over Z2 [3, 32], recovery in the stochastic block model [2, 7], and Synchronization over Z2 [1, 7, 18]. When 𝒢SO(2) and the functions fij are linear with respect to the representation ρ1:SO(2)C given by ρ1(θ)=eiθ, it coincides with the semidefinite relaxation for angular synchronization [40]. Similarly, when 𝒢O(d) and the functions are linear with respect to the natural d-dimensional representation, then the NUG problem essentially coincides with the little Grothendieck problem over the orthogonal group [9, 31]. Other examples include the shape matching problem in computer graphics for which 𝒢 is the permutation group (see [24, 16]). In addition, it has been shown in [13] that the formulation of NUG and the algorithms presented in this paper can be extended to simultaneous alignment and classification of a mixture of different signals.

1.1. Orientation estimation in cryo-Electron Microscopy.

A particularly important application of this framework is the orientation estimation problem in cryo-Electron Microscopy [39].

Cryo-EM is a technique used to determine the 3-dimensional structure of biological macromolecules. The molecules are rapidly frozen in a thin layer of ice and imaged with an electron microscope, which gives noisy 2-dimensional projections. One of the main difficulties with this imaging process is that these molecules are imaged at different unknown orientations in the sheet of ice and each molecule can only be imaged once (due to the destructive nature of the imaging process). More precisely, each measurement consists of a tomographic projection of a rotated (by an unknown rotation) copy of the molecule. The task is then to reconstruct the molecule density from many such noisy measurements. Although in principle it is possible to reconstruct the 3-dimensional density directly from the noisy images without estimation of the rotations [25], or by treating rotations as nuisance parameters [47, 6] here we consider the problem of estimating the rotations directly from the noisy images. In Section 2, we describe how this problem can be formulated in the form (1.1).

2. Multireference Alignment

In classical linear inverse problems, one is tasked with recovering an unknown element x𝒳 from a noisy measurement of the form 𝒫(x)+ϵ, where ϵ represents the measurement error and 𝒫 is a linear observation operator. There are, however, many problems where an additional difficulty is present; one class of such problems includes non-linear inverse problems in which an unknown transformation acts on x prior to the linear measurement. Specifically, let 𝒳 be a vector space and 𝒢 be a group acting on 𝒳. Suppose we have n measurements of the form

yi=𝒫gix+ϵi,i=1,,n (2.1)

where

  • x is a fixed but unknown element of 𝒳,

  • g1,,gn are unknown elements of 𝒢,

  • is the action of 𝒢 on 𝒳,

  • 𝒫:𝒳Y is a linear operator,

  • Y is the (finite-dimensional) measurement space,

  • ϵi’s are independent noise terms.

If the gi’s were known, then the task of recovering x would reduce to a classical linear inverse problem, for which many effective techniques exist. While in many situations it is possible to estimate x directly without estimating g1,,gn, or by treating these as nuisance parameters, here we focus on the problem of estimating the group elements g1,,gn.

There are several common approaches for inverse problems of the form (2.1). One is motivated by the observation that estimating x knowing the gi’s and estimating the gi’s knowing x are both considerably easier tasks. This suggests an alternating minimization approach where each estimation is updated iteratively. Besides a lack of theoretical guarantees, convergence may also depend on the initial guess. Another approach, which we refer to as pairwise comparisons [40], consists in determining, from pairs of observations yi,yj, the most likely value for gigj-1. Although the problem of estimating the gi’s from these pairwise guesses is fairly well-understood [40, 10, 43] enjoying efficient algorithms and performance guarantees, this method suffers from loss of information as not all of the information of the problem is captured in this most likely value for gigj-1 and thus this approach tends to fail at low signal-to-noise-ratio.

In contrast, the Maximum Likelihood Estimator (MLE) leverages all information. Assuming that the ϵi’s are i.i.d. Gaussian, the MLE for the observation model (2.1) is given by the following optimization problem:

minimizeg1,,gn,xi=1nyi𝒫gix22subject togi𝒢x𝒳 (2.2)

We refer to (2.2) as the Multireference Alignment (MRA) problem. Let us denote the ground truth signal and group elements by x0 and g10,,gn0; the solution to the optimization problem by x* and g1*,,gn*, which we will also refer to as xMLE and g1MLE,,gnMLE. Unfortunately, the exponentially large search space and nonconvex nature of (2.2) often render it computationally intractable. However, for several problems of interest, we formulate (2.2) as an instance of an NUG for which we develop computationally tractable approximations.

Notice that although MLE typically enjoys several theoretical properties, their underlying technical conditions do not hold in this case. Specifically, the number of parameters to be estimated is not fixed but rather grows indefinitely with the sample size n: for each sample yi there is a group element gi that needs to be estimated. As a result, the MLE may not be consistent in this case. In other words, even in the limit n the estimator xMLE may not converge to the ground truth x0. Similarly, the estimated group elements will not converge to their true values. A different version of MLE, not considered in this paper, in which the group elements are treated as nuisance parameters and are marginalized would enjoy the nice theoretical properties.

2.1. Registration of signals on the sphere.

Consider the problem of estimating a bandlimited signal on the circle x:S1C from noisy rotated discrete sampled copies of it. In this problem, 𝒳=spaneikθk=-tt is the space of bandlimited functions up to degree t on S1,𝒢=SO(2) and the group action is

gx=k=-ttαkeikθ-θg,

where x=k=-ttαkeikθ and we identified gSO(2) with θg[0,2π].

The measurements are of the form

yi𝒫gix+ϵi,i=1,,n

where

  • x𝒳,

  • giSO(2),

  • 𝒫:𝒳CL samples the function at L equally spaced points in S1,

  • ϵi~𝒩0,σ2IL×L(i=1,,n) are independent Gaussians.

Our objective is to estimate g1,,gn and x. Since estimating x knowing the group elements gi is considerably easier, we will focus on estimating g1,,gn. As shown below, this will essentially reduce to the problem of aligning (or registering) the observations y1,,yn.

In absence of noise, the problem of finding the gi’s is trivial (cf. first column of Figure 2.1). With noise, if x is known (as it is in some applications), then the problem of determining the gi’s can be solved by matched filtering (cf. second column of Figure 2.1). However, x is unknown in general. This, together with the high levels of noise, render the problem significantly more difficult (cf. last two columns of Figure 2.1).

Figure 2.1.

Figure 2.1.

Illustration of the registration problem in S1. The first column consists of a noiseless signal at three different shifts, the second column represents an instance for which the template x is known and matched filtering is effective to estimate the shifts. However, in the examples we are interested in the template is unknown (last two columns) rendering the problem of estimating the shifts significantly harder.

We now define the problem of registration in d-dimensions in general. 𝒳=spanpkk𝒜t is the space of bandlimited functions up to degree t on Sd where the pk’s are orthonormal polynomials on Sd,𝒜t indexes all pk up to degree t and 𝒢=SO(d+1).

The measurements are of the form

yi𝒫gix+ϵi,i=1,,n (2.3)

where

  • x𝒳,

  • giSO(d+1),

  • 𝒫:𝒳CL samples the function on L points in Sd,

  • ϵi~𝒩0,σ2IL×L(i=1,,n) are independent Gaussians.

Again, our objective is to estimate g1,,gn and x. We would like the sampling operator 𝒫 to be ‘uniform’. One possible sampling scheme is spherical designs surveyed in [11]. An illustration of signals on a sphere, sampled at such points, is provided in Figure 2.2.

Figure 2.2.

Figure 2.2.

An illustration of registration in 2-dimensions. The left four spheres provide examples of clean signals yi and the right four spheres are of noisy observations. Note that the images are generated using a quantization of the sphere.

The MRA solution for registration in d-dimensions is given by

minimizeg1,,gn,xi=1nyi𝒫gix22 subject togiSOd+1x𝒳 (2.4)

We now remove x from (2.4). Let 𝒬:CL𝒳 be the adjoint of 𝒫.𝒬 is also an approximate inverse of 𝒫 (up to normalization), because points are sampled from a t-design which has the property of exactly integrating polynomials on the sphere. Then, yi-𝒫gix22𝒬yi-gix22 (up to normalization), and the approximation error decreases as L increases. Since gi preserves the norm 2, it follows that (2.4) is equivalent to

minimizeg1,,gn,xi=1ngi-1𝒬yi-x22subject togiSO(d+1)x𝒳. (2.5)

Since the minimizer x with fixed gi’s is the average, (2.5) is equivalent to

minimizeg1,,gni,j=1ngi1𝒬(yi)gj1𝒬(yj)22 subject togiSO(d+1). (2.6)

Since gi preserves 2 norm, then (2.6) is equivalent to

minimizeg1,,gni,j=1n𝒬(yi)gigj1𝒬(yj)22 subject togiSO(d+1). (2.7)

In summary, (2.4) can be approximated by (2.7), which is an instance of (1.1).

2.2. Orientation estimation in cryo-EM.

The task here is to reconstruct the molecule density from many noisy tomographic projection images (see the right column of Figure 1.1 for an idealized density and measurement dataset). We assume the molecule does not have any non-trivial point group symmetry. The linear inverse problem of recovering the molecule density given the rotations fits in the framework of classical computerized tomography for which effective methods exist. Thus, we focus on the non-linear inverse problem of estimating the unknown rotations and the underlying density.

Figure 1.1.

Figure 1.1.

Illustration of the cryo-EM imaging process: A molecule is imaged after being frozen at a random (unknown) rotation and a tomographic 2-dimensional projection is captured. Given a number of tomographic projections taken at unknown rotations, we are interested in determining such rotations with the objective of reconstructing the molecule density. Images courtesy of Amit Singer and Yoel Shkolnisky [42].

An added difficulty is the high level of noise in the images. In fact, it is already non-trivial to distinguish whether a molecule is present in an image or if the image consists only of noise (see Figure 2.3 for a subset of an experimental dataset). On the other hand, these datasets consist of many projection images which renders reconstruction possible.

Figure 2.3.

Figure 2.3.

Sample images from the E. coli 50S ribosomal subunit, generously provided by Dr. Fred Sigworth at the Yale Medical School.

We formulate the problem of orientation estimation in cryo-EM. Let 𝒳 be the space of bandlimited functions that are also essentially compactly supported in R3 and 𝒢=SO(3). For perfectly centered images, and ignoring the effect of the microscope’s contrast transfer function, the measurements are of the form

Ii(x,y)𝒫giϕ+ϵi,i=1,,n (2.8)
  • ϕ𝒳,

  • giSO(3),

  • 𝒫(ϕ) samples -ϕ(x,y,z)dz(𝒫 is called the discrete X-ray transform),

  • ϵi’s are i.i.d Gaussians representing noise.

Our objective is to find g1,,gn and ϕ.

The operator 𝒫 in the orientation estimation problem is different than in the registration problem. Specifically, 𝒫 is a composition of tomographic projection and sampling. To write the objective function for the orientation estimation problem, we will use the Fourier slice theorem [30].

The Fourier slice theorem states that the 2-dimensional Fourier transform of a tomographic projection of a molecular density ϕ coincides with the restriction to a plane normal to the projection direction, a slice, of the 3-dimensional Fourier transform of the density ϕ. See Figure 2.4.

Figure 2.4.

Figure 2.4.

An illustration of the use of the Fourier slice theorem and the common lines approach to the orientation estimation problem in cryo-EM. Image courtesy of Amit Singer and Yoel Shkolnisky [42].

Let I^i(r,θ) be the Fourier transform of Ii in polar coordinates. We identify Iˆi and Iˆj with the xy-plane in R3, and apply gi-1 and gj-1 to Iˆi and Iˆj, respectively. Then, the directions of the lines of intersection on Iˆi and Iˆj are given, respectively, by unit vectors

cij(gigj1)=gi(gi1e3×gj1e3)gi(gi1e3×gj1e32=e3×gigj1e3e3×gigj1e32, (2.9)
cji(gigj1)=gj(gi1e3×gj1e3)gj(gi1e3×gj1e3))2=(gigj1)1e3×e3(gigj1)1e3×e32. (2.10)

where e3(0,0,1)T. See [42] for details.

Since the noiseless images should agree on their common lines, we consider the following MRA-like cost function:

minimizeg1,,gni,j=1nI^i(,cij(gigj1))I^j(,cji(gigj1))22 subject to giSO(3), (2.11)

where, with a minor abuse of notation, we identify the vector cij with the angle θij of a common line in the Fourier transform of an image Iˆi. Equation (2.11) is an instance of (1.1). Note that we could also use the L1 norm or a weighted L2 norm in the cost function.

Note that for n=2 images, there is always a degree of freedom along the line of intersection. In other words, we cannot recover the true orientation between Iˆ1 and Iˆ2. However, for n3, this degree of freedom is eliminated. It is also worth mentioning several important references in the context of angular reconstitution [22, 45]. In general, the measurement system suffers from a handedness ambiguity on the reconstruction (see, for example, [42]), this will be discussed in detail later in the paper.

3. Linearization via Fourier expansion

Let us consider the objective function in the general form

i,j=1nfij(gigj1). (3.1)

Note that each fij in (3.1) can be nonlinear and nonconvex. However, since 𝒢 is compact (and since fijC(𝒢)), we can expand, each fij in Fourier series. More precisely, given the unitary irreducible representations ρk of 𝒢, we can write

fij(gigj1)=k=0dktr[f^ij(k)ρk(gigj1)]=k=0dktr[f^ij(k)ρk(gi)ρk*(gj)], (3.2)

where fˆij(k) are the Fourier coefficients of fij and can be computed from fij via the Fourier transform

f^ij(k)𝒢fij(g)ρk(g1)dg=𝒢fij(g)ρk*(g)dg. (3.3)

Above, dg denotes the Haar measure on 𝒢 and dk the dimension of the representation ρk. See [17] for an introduction to the representations of compact groups.

We express the objective function (3.1) as

i,j=1nfij(gigj1)=i,j=1nk=0dktr[f^ij(k)ρk(gi)ρk*(gj)]=k=0i,j=1ndktr[f^ij(k)ρk(gi)ρk*(gj)],

which is linear in ρkgiρk*gj. This motivates writing (1.1) as linear optimization over the variables

X(k)ρkg1ρkgnρkg1ρkgn*.

In other words,

i,j=1nfijgigj-1=k=0trCkXk,

where the coefficient matrices are given by

C(k)dk[f^11(k)f^21(k)f^n1(k)f^12(k)f^22(k)f^n2(k)f^1n(k)f^2n(k)f^nn(k)].

We refer to the dk×dk block of X(k) corresponding to ρkgiρk*gj=ρkgigj-1 as Xij(k). We now turn our attention to constraints on the variables X(k)k=0. It is easy to see that:

X(k)0,k (3.4)
Xii(k)=Idk×dk,k,i, (3.5)
rankX(k)=dk,k, (3.6)
Xij(k)Imρk,k,i,j. (3.7)

Constraints (3.4), (3.5) and (3.6) ensure X(k) is of the form

X(k)=X1(k)X2(k)Xn(k)X1(k)X2(k)Xn(k)*,

for some Xi(k) unitary dk×dk matrices. The constraint (3.7) attempts to ensure that Xi(k) is in the image of the representation of 𝒢. Notably, none of these constraints ensures that, for different values of k,Xij(k) correspond to the same group element gigj-1. Adding such a constraint would yield

minimizeX(k)k=0trC(k)X(k)subject toX(k)0Xii(k)=Idk×dkrankX(k)=dkXij(k)=ρkgigj-1, (3.8)

where gi and gj are elements of 𝒢.

Unfortunately, both the rank constraint and the last constraint in (3.8) are, in general, nonconvex. We will relax (3.8) by dropping the rank requirement and replacing the last constraint by positivity constraints that couple different X(k)’s. We achieve this by considering the Dirac delta funcion on 𝒢. Notice that the Dirac delta funcion δ(g) on the identity e𝒢 can be expanded as

δ(g)=k=0dktr[δ^(k)ρk(g)]=k=0dktr[(𝒢δ(h)ρk*(h)dh)ρk(g)]=k=0dktr[ρk(g)].

If we replace g with g-1gigj-1, then we get

δ(g1gigj1)=k=0dktr[ρk(g1)ρk(gigj1)]=k=0dktr[ρk*(g)Xij(k)].

To arrive at a convex program, we consider the following convex constraints, that form a natural convex relaxation for Dirac deltas,

k=0dktr[ρk*(g)Xij(k)]0g𝒢, (3.9)
𝒢(k=0dktr[ρk*(g)Xij(k)])dg=1. (3.10)

This suggests relaxing (3.8) to

minimizeX(k)k=0trC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=0dktrρk*(g)Xij(k)0g𝒢𝒢k=0dktrρk*(g)Xij(k)dg=1. (3.11)

For a nontrivial irreducible representation ρk, we have 𝒢ρk(g)dg=0. This means that the integral constraint in (3.11) is equivalent to the constraint

Xij(0)=1,i,j.

Thus, we focus on the optimization problem

minimizeX(k)k=0trC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=0dktrρk*(g)Xij(k)0g𝒢Xij(0)=1. (3.12)

When 𝒢 is a finite group it has only a finite number of irreducible representations. This means that (3.12) is a semidefinite program and can be solved, to arbitrary precision, in polynomial time [46]. In fact, when 𝒢ZL, a suitable change of basis shows that (3.12) is equivalent to the semidefinite programming relaxation proposed in [8] for the signal alignment problem.

Unfortunately, many of the applications of interest involve infinite groups. This creates two obstacles to solving (3.12). One is due to the infinite sum in the objective function and the other due to the infinite number of positivity constraints. In the next section, we address these two obstacles for the groups SO(2) and SO(3).

4. Finite truncations for SO(2)andSO(3) via Fejér kernels

The objective of this section is to replace (3.12) by an optimization problem depending only in finitely many variables X(k). The objective function in (3.12) is converted from an infinite sum to a finite sum by truncating at degree t. That is, we fix a t and set C(k)=0 for k>t. This consists of truncating the Fourier series of i,j=1nfijgigj-1. Unfortunately, constraint (3.9) given by

k=0dktrρk*gXijk0g𝒢,

still involves infinitely many variables Xij(k) and consists of infinitely many linear constraints.

We now address this issue for the groups SO(2) and SO(3).

4.1. Truncation for SO(2).

Since we truncated the objective function at degree t, it is then natural to truncate the infinite sum in constraint (3.9) also at t. If we truncated below t, then some variables (such as X(t)) are not constrained; and if we truncated above t, then some variables (such as X(t+1)) do not affect the cost function. The irreducible representations of SO(2) are eikθ, and dk=1 for all k. Let us identify gSO(2) with θg[0,2π]. That straightforward truncation corresponds to approximating the Dirac delta with

δgk=-tteikθg.

This approximation is known as the Dirichlet kernel, which we denote as

Dtθk=-tteikθ.

However, the Dirichlet kernel does not inherit all the desirable properties of the delta function. In fact, Dt(θ) is negative for some values of θ.

Instead, we use the Fejér kernel, which is a non-negative kernel, to approximate the Dirac delta. The Fejér kernel is defined as

Ftθ1tk=0t-1Dk=k=-tt1-kteikθ,

which is the first-order Cesàro mean of the Dirichlet kernel.

This motivates us to replace constraint (3.9) with

k=-tt1-|k|te-ikθXij(k)0θ0,2π,

where, for k>0,Xij(-k) denotes Xij(k)*.

This suggests considering

minimizeX(k)k=0ttrC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=-tt1-|k|te-ikθXij(k)0θ[0,2π]Xij(0)=1,

which only depends on the variables Xij(k) for k=0,,t.

Unfortunately, the condition that the trigonometric polynomial k=-tt1-|k|te-ikθXij(k) is always non-negative, still involves an infinite number of linear in-equalities. Interestingly, due to the Fejér-Riesz factorization theorem (see [19]), this condition can be replaced by an equivalent condition involving a positive semidefinite matrix — it turns out that every nonnegative trigonometric polynomial is a square, meaning that the so called sum-of-squares relaxation [33, 34] is exact. However, while such a formulation would still be an SDP and thus solvable, up to arbitrary precision, in polynomial time, it would involve a positive semidefinite variable for every pair (i,j), rendering it computationally challenging. For this reason we relax the non-negativity constraint by asking that k=-tt1-|k|te-ikθXij(k) is non-negative in a finite set ΩtSO(2). This yields the following optimization problem:

minimizeX(k)k=0ttrC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=-tt1-|k|te-ikθXij(k)0θΩtXij(0)=1. (4.1)

4.2. Truncation for SO(3).

The irreducible representations of SO(3) are the Wigner-D matrices W(k)(α,β,γ), and dk=2k+1. See [48] for an introduction to Wigner-D matrices. Let us associate gSO(3) with Euler (Z-Y-Z) angle (α,β,γ)[0,2π]×[0,π]×[0,2π]. A straightforward truncation yields the approximation

δgk=0t2k+1trWkα,β,γ.

Observe that the operator tr is invariant under conjugation. Then W(k) can be decomposed as

W(k)(α,β,γ)=RΛ(k)(θ)R*

with an R such that

Λkθ=e-ikθ1eikθ.

We can think of R as a change of basis and Λ(k)(θ) as a rotation from SO(2) under the basis R. It follows that

trWkα,β,γ=trΛkθ=m=-kkeimθ=Dkθ.

The relationship between θ and α,β,γ is

θ=2arccoscosβ2cosα+γ2.

This relationship can be obtained by directly evaluating trW(1)(α,β,γ) using the Wigner-d matrix w(1):

tr[W(1)(α,β,γ)]=m=11Wm,m(1)(α,β,γ)=m=11eim(α+γ)wm,m(1)(β)=cos(β)(1+cos(α+γ))+cos(α+γ).

See [48] also for an introduction to Wigner-d matrix. This straightforward truncation at t yields

δgk=0t2k+1Dkθ,

which, again, inherits the undesirable property that this approximation can be negative for some θ. Recall that we circumvented this property in the 1-dimension case by taking the first-order Cesàro mean of the Dirichlet kernel. In the 2-dimension case, we will need the second-order Cesàro mean. Notice that

Dkθ=sin2k+1θ2sinθ2.

Fejér proved that [4]

k=0t(3)t-k(t-k)!k+12sin(2k+1)θ20,0θπ

where (3)t-k(t-k)!=12(t-k+2)(t-k+1). It follows that

k=0t(3)tk(tk)!(k+12)Dk(θ)=k=0t(3)tk(tk)!(k+12)sin[(2k+1)θ2]sin(θ2)0,π0π.

Let us define

Ft(g)=Ft(α,β,γ)k=0t(3)t-k(t-k)!k+12sin(2k+1)θg2sinθg2

where θg=2arccoscosβ2cosα+γ2.

We replace constraint (3.9) with

Ftα,β,γ0α,β,γ0,2π×0,π×0,2π.

Secondly, we discretize the group SO(3) to obtain a finite number of constraints. We consider a suitable finite subset ΩtSO(3). In our implementation, we use a Hopf fibration [49] to discretize SO(3). The quotient space SO(3)/SO(2) is equivalent to S2. We take a uniform discretization of S1SO(2) and a spherical design [11] of S2. It is possible to find a spherical design on S2 with 𝒪r2 points [11]. By [49], we use 𝒪(r) points to discretize SO(2). The size of our SO(3) discretization is 𝒪r3. So, we have to enforce 𝒪r3 inequality constraints. The choice of r is up to the user to strike a balance between computational speed and accuracy. We can then relax the non-negativity constraint yielding the following semidefinite program1:

minimizeX(k)k=0ttrC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=0t(3)t-k(t-k)!k+12trW(k)(α,β,γ)*Xij(k)0(α,β,γ)ΩtXij(0)=1. (4.2)

4.3. An additional constraint on X(1).

In this section, we discuss an additional constraint on X(1), which uses properties of quaternions to constrain each block Xij(1) of X(1) in the convex hull of SO(3) more directly.

We consider the standard rotation matrix R and the unit quaternion q=qr+qii+qjj+qkk which represent the same rotation as the block Xij(1) (whose representation is associated with spherical harmonics), and we consider the outer product Qij of the unit quaternion q

Qij=qTq. (4.3)

The standard rotation matrix R is related to the block Xij(1) by the formula

R*=M*T*Xij(1)TM,

where,

T-i2012010-i20-12,

and

M010001-100.

Indeed, one can verify that

T*Xij(1)T=R[22]R[32]-R[12]R[23]R[33]-R[13]-R[21]-R[31]R[11],

where,

R=R[11]R[12]R[13]R[21]R[22]R[23]R[31]R[32]R[33],

and M is used simply to rearrange the elements of T*Xij(1)T.

Next, R is mapped to Qij by Rodrigues’ rotation formula (see [37]):

Qij=R141+R11+R22+R33R32-R23R13-R31R21-R12R32-R231+R11-R22-R33R21+R12R31+R13R13-R31R21+R121+R22-R11-R33R23+R32R21-R12R31+R13R23+R321+R33-R11-R22.

In summary, the mapping from Xij(1) to Qij, which we denote by 𝒜Eq, is given by the formula:

Qij=𝒜EqXij(1)=M*T*Xij(1)TM.

For each Qij, we wish to impose the constraints implied by Equation (4.3):

Qij0,trQij=1,rankQij=1.

But since the constraint rankQij=1 is not convex, we will drop it, giving us the following SDP in place of (4.2):

minimizeX(k)k=0ttrC(k)X(k)subject toX(k)0Xii(k)=Idk×dkk=0t(3)t-k(t-k)!k+12trW(k)(α,β,γ)*Xij(k)0(α,β,γ)ΩtXij(0)=1Qij=𝒜EqXij(1),Qij0,trQij=1. (4.4)

5. Applications

In this section, we consider the application of (4.1) to the problem of registration over the unit circle and the application of (4.4) to registration over the unit sphere and orientation estimation in cryo-EM. To solve the SDP for each problem, the only parameters we need to determine are the coefficient matrices C(k) and the truncation parameter t. The calculations of C(k) are detailed in the following pages. As for t, we experimented over a range of values for each problem, and chose a value that balances computational time with accuracy in the estimated rotations. The SDP outputs the relative transformations gij=gigj-1, while we need g1,,gn. For each problem, we describe a rounding procedure to recover g1,,gn from the gij’s.

5.1. Registration in 1-dimension.

Recall that 𝒳 is the space of bandlimited functions up to degree t on S1. That is, for x𝒳, we can express

xω=l=-ttαleilω.

Again, the irreducible representations of SO(2) are eikθ, and dk=1 for all k. Let us identify gSO(2) with θg[0,2π], then

gxω=l=-tteilθgαleilω.

Let 𝒫 sample the underlying signal x at L=2t+1 distinct points. This way, we can determine all the αl’s associated with x.

Since yi=𝒫gix+ϵi, for the adjoint 𝒬, we have

𝒬yiω=l=-ttαlieilω.

Let us identify gigj-1 with θij[0,2π]. Then, we can express fij in terms of αl(i),αl(j) and θij:

fij(gigj1)=𝒬(yi)gigj1𝒬(yj)22=S1|l=tt(αl(i)αl(j)eilθij)eilω|2dω=l=tt|αl(i)αl(j)eilθij|2.

The Fourier coefficients of fij are

f^ij(k)=02πl=tt|αl(i)αl(j)eilθij|2eikθijdθij=02πl=tt(|αl(i)|2eikθij+|αl(j)|2eikθijαl(i)αl(j)¯ei(kl)θijαl(i)¯αl(j)ei(k+l)θij)dθij=2π{l=tt(|αl(i)|2+|αl(j)|2)α0(i)α0(j)¯α0(i)¯α0(j),k=0αk(i)αk(j)¯αk(i)¯αk(j),k0

Note that we re-indexed the coefficients fˆij(k)fˆij(k-(t+1)).

5.1.1. Rounding.

(4.1) gives us the X(k)’s. From the X(k)’s, we want to extract each θi[0,2π] (up to a global transformation). Let us consider X(1). We want X(1) to be of the form

eiθ1eiθ2eiθneiθ1eiθ2eiθn*.

Although X(1) is not guaranteed to be rank 1, we will simply take the top eigenvector of X(1) as our estimate of eiθ1,,eiθn. And from eiθ1,,eiθn, we can recover θ1,,θn. See [40] for the reasoning behind this approach. In practice, we find using the top eigenvector of X(1) is a sufficient estimate of θ1,,θn. We do not need to run additional comparisons against X(2),,X(t).

5.2. Registration in 2-dimension.

Recall that 𝒳 is the space of bandlimited functions up to degree t on S2. That is, for x𝒳, we can express

xω=l=0tm=-llαl,mYl,mω,

where Ylm are the spherical harmonics. Again, the irreducible representations of SO(3) are the Wigner D-matrices W(k)(α,β,γ), and dk=2k+1. Let us associate gSO(3) with Euler (Z-Y-Z) angle (α,β,γ)[0,2π]×[0,π]×[0,2π], then

gxω=l=0tm,m=-llWm,mlα,β,γαl,mYl,mω.

Let 𝒫 sample the underlying signal x at L=(t+1)2 points. This way, we can determine all the αlm’s associated with x.

Again, for the adjoint 𝒬 we have

𝒬yiω=l=0tm=-llαl,miYl,mω.

Let us identify gigj-1SO(3) with Euler (Z-Y-Z) angle αij,βij,γij[0,2π]×[0,π]×[0,2π]. Then, we can express fij in terms of αl,m(i),αl,m(j) and αij,βij,γij:

fij(gigj1)=𝒬(yi)gigj1𝒬(yj)22=l=0tm=ll(|αl,m(i)|2+|αl,m(j)|2)l=0tm,m=llαl,m(i)Wm,m(l)¯(αij,βij,γij)αl,m(j)¯l=0tm,m=llαl,m(i)¯Wm,m(l)(αij,βij,γij)αl,m(j).

The Fourier coefficients are given by

f^ij(k)=SO(3)fij(g)(W(k)(α,β,γ))*dg=8π22k+1{l=0tm=ll(|αlm(i)|2+|αlm(j)|2)α00(i)α00(j)¯α00(i)¯α00(j),k=0((1)mmαk,m(i)αk,m(j)¯αk,m(i)¯αk,m(j))m,m=kk,k0

Here, we used the orthogonality relationship

SO3Wm1,m2k-α,β,γWm1,m2kα,β,γdg=8π22k+1δk,kδm1,m1δm2,m2,

and the property

Wm,mk(α,β,γ)=(-1)m-mW-m,-mk-α,β,γ.

5.2.1. Rounding.

Again, (4.4) gives us the X(k)’s. From the X(k)’s, we want to extract each (α,β,γ)[0,2π]×[0,π]×[0,2π] (up to a global transformation). Let us consider X(1). We want X(1) to be of the form

W(1)α1,β1,γ1W(1)α2,β2,γ2W(1)αn,βn,γnW(1)α1,β1,γ1W(1)α2,β2,γ2W(1)αn,βn,γn*,

where each W(1)αi,βi,γi is a 3 × 3 matrix. Similarly, X(1) is not guaranteed to be rank 3, but we will simply take the top 3 eigenvector of X(1) as our estimate of W(1)α1,β1,γ1,,W(1)αn,βn,γn. And from W(1)α1,β1,γ1,,W(1)αn,βn,γn, we can recover α1,β1,γ1,,αn,βn,γn.

5.3. Orientation estimation in cryo-EM.

We refer to [50] to expand the objective function. We emphasize that the theory holds for arbitrary basis on the space containing the Iˆi’s. We choose to construct the C(k)’s using coefficients and parameters from the Fourier-Bessel expansion. Projection Iˆi can be expanded via Fourier-Bessel series as

I^i(r,θ)=k=q=1αkq(i)ψkq(c)(r,θ),

where

ψkq(c)(r,θ)=NkqJkRkqrceikθ,rc,0,r>c.

The parameters above are defined as follows:

  • c is the radius of the disc containing the support of Iˆi,

  • Jk is the Bessel function of integer order k,

  • Rkq is the qth root of Jk,

  • Nkq=1cπJk+1Rkq is a normalization factor.

To avoid aliasing, we truncate the Fourier-Bessel expansion as follows.

I^i(r,θ)k=kmaxkmaxq=1pkαkq(i)ψkq(c)(r,θ).

See [50] for a discussion on kmax and pk. For the purpose of this section, let us assume we have αkq(i):-kmaxkkmax,1qpk for each Iˆi. (These can be computed from the Cartesian grid sampled images.)

We shall determine the relationship between Iˆir,θi and Iˆjr,θj, and the lines of intersection between gi-1Iˆi and gj-1Iˆj embedded in R3. Recall from (2.9) and (2.10) that the directions of the lines of intersection between gi-1Iˆi and gj-1Iˆj are given, respectively, by unit vectors

cij(gigj1)=e3×gigj1e3e3×gigj1e32,
cji(gigj1)=(gigj1)1e3×e3(gigj1)1e3×e32.

Let us associate gigj-1SO(3) with Euler (Z-Y-Z) angle αij,βij,γij[0,2π]×[0,π]×[0,2π]. Then

e3×gigj1e3=[sinγijsinβijcosγijsinβij0],
(gigj1)1e3×e3=[sinαijsinβijcosαijsinβij0],

under the rotation matrix RZγijRYβijRZαij. The directions of the lines of intersection in Iˆi and Iˆj under gigj-1 are in the directions, respectively,

θi=arctansinγij,-cosγij=γij-π2,
θj=arctan-sinαij,-cosαij=-αij-π2.

We express the fij’s in terms of αkq(i),αkq(j), and θi and θj:

fij(θi,θj)fij(gigj1)=k=kmaxkmaxq=1pk(αkq(i)ψkq(c)(r,θi)αkq(j)ψkq(c)(r,θj))L22=k,k,q,qcNkqNkq(αkq(i)eikθiαkq(j)eikθj)(αkq(i)eikθiαkq(j)eikθj)*01Jk(Rkqr)Jk(Rkqr)dr.

For each k,k,q,q, we approximate the integral

01JkRkqrJkRkqrdr

with a Gaussian quadrature.

Using the approximation above, we have

fij(θi,θj)k,q,k,qbk,q,k,q(αkq(i)αkq(i)¯ei(kk)θi+αkq(j)αkq(j)¯ei(kk)θjαkq(i)αkq(j)¯ei(kθikθj)αkq(j)αkq(i)¯ei(kθjkθi)),

where

bk,q,k,q=cNkqNkqiwiJkRkqriJkRkqri.

In terms of the Euler (Z-Y-Z) angles,

fij(αij,γij)fij(θi,θj)k,q,k,qbk,q,k,qeiπ2(kk)(αkq(i)αkq(i)¯ei(kk)γij+αkq(j)αkq(j)¯ei(kk)αijαkq(i)αkq(j)¯eikγij+ikαijαkq(j)αkq(i)¯eikαijikγij).

The Fourier coefficients are given by

f^ij(k)=SO(3)fij(α,γ)(W(k)(α,β,γ))*dg=02π02πfij(α,γ)(0πW(k)(γ+π,β,α+π)sinβdβ)dαdγ.

Note that

0πW(k)(α,β,γ)sinβdβm,m=0πeimαwm,m(k)(β)eimγsinβdβ,

where w(k) is the Wigner d-matrix. Let us define

Hkm,m(-1)m+m0πwm,m(k)(β)sinβdβ=(-1)m+m0πim-ml=-kkwl,m(k)(π/2)e-ilβwl,m(k)(π/2)sinβdβ=(-1)m+mim-m2w0,m(k)(π/2)w0,m(k)(π/2)-iπ2w1,m(k)(π/2)w1,m(k)(π/2)+iπ2w-1,m(k)(π/2)w-1,m(k)(π/2)+|l|2k1+e-ilπ1-l2wl,m(k)(π/2)wl,m(k)(π/2).

The m,mth entry of fˆij(k) is approximated by

(f^ij(k))m,m=Hk(m,m)02π02πfij(α,γ)eimγeimαdαdγ=4π2Hk(m,m)k,q,k,qbk,q,k,q(αkq(i)αkq(i)¯δ{m=kk}δ{m=0}+αkq(j)αkq(j)¯δ{m=0}δ{m=kk}αkq(i)αkq(j)¯δ{m=k}δ{m=k}αkq(j)αkq(i)¯δ{m=k}δ{m=k}).

Here, δ is the Kronecker delta and bk,q,k,q absorbed e-iπ2k-k.

5.3.1. Handedness ambiguity in cryo-EM.

There exists one additional issue specifically for the cryo-EM problem arising from the handedness ambiguity. Suppose an image was projected from some molecular density ϕ and orientation R. Let J be the reflection operator across the imaging plane. The molecular density Jϕ under orientation JRJ* would produce the same projection. In other words, the set of projection images can belong to different molecular densities ϕ and Jϕ under different rotations.

We will formalize and deal with the handedness ambiguity in terms of the Wigner D-matrices. Recall that the Wigner D-matrix W(k)(g) corresponding to (α,β,γ)SO(3) is

Wkg=eimαwm,mkβeimγm,m=-kk.

Let J(k) be the following (2k+1)×(2k+1) diagonal matrix:

J(k)-11-1.

(The diagonal alternates between +1 and −1.) Due to the handedness ambiguity, if W(k)giW(k)gj-1k is a solution to (4.4), then J(k)W(k)giW(k)gj-1J(k)k is also a valid solution to (4.4). In fact, for any h[0,1],

hW(k)giW(k)gj-1+(1-h)J(k)W(k)giW(k)gj-1J(k)k

is a valid solution to (4.4).

Let us remove this extra degree of freedom h. Observe that for h=12,

12Wkg+12JkWkgJk=eimαwm,mkβeimγ,m+m0mod2,0,otherwise.

I.e., the odd-indexed entries are 0. For example, in the case of k=1,

12W(1)(g)+12J(1)W(1)(g)J(1)=[eiαw1,1(1)(β)eiγ0eiαw1,1(1)(β)eiγ0w0,0(1)(β)0eiαw1,1(1)(β)eiγ0eiαw1,1(1)(β)eiγ],

and in the case of k=2,

12W(2)(g)+12J(2)W(2)(g)J(2)=e-2iαw-2,-2(2)(β)e-2iγ0e-2iαw-2,0(2)(β)0e-2iαw-2,2(2)(β)e2iγ0e-iαw-1,-1(2)(β)e-iγ0e-iαw-1,1(2)(β)eiγ0w0,-2(2)(β)e-2iγ0w0,0(2)(β)0w0,2(2)(β)e2iγ0eiαw1,-1(2)(β)e-iγ0eiαw1,1(2)(β)eiγ0e2iαw2,-2(2)(β)e-2iγ0e2iαw2,0(2)(β)0e2iαw2,2(2)(β)e2iγ.

We constrain the odd-indexed entries of Xij(k)’s to be 0 so that the SDP finds the solution with h=12. Note that, in practice, we do not explicitly add this constraint. Instead, we permute Xij(k) into two disjoint diagonal blocks. For example, in the case of k=1,

12W(1)(g)+12J(1)W(1)(g)J(1)=[w0,0(1)(β)000eiαw1,1(1)(β)eiγeiαw1,1(1)(β)eiγ0eiαw1,1(1)(β)eiγeiαw1,1(1)(β)eiγ],

and in the case of k=2,

12W(2)(g)+12J(2)W(2)(g)J(2)=[eiαw1,1(2)(β)eiγeiαw1,1(2)(β)eiγ000eiαw1,1(2)(β)eiγeiαw1,1(2)(β)eiγ00000e2iαw2,2(2)(β)e2iγe2iαw2,0(2)(β)e2iαw2,2(2)(β)e2iγ00w0,2(2)(β)e2iγw0,0(2)(β)w0,2(2)(β)e2iγ00e2iαw2,2(2)(β)e2iγe2iαw2,0(2)(β)e2iαw2,2(2)(β)e2iγ].

We can conjugate each Xij(k) in (4.4) by a permutation and get

Xij(k)=Xij(k,0)00Xij(k,1).

Similarly, we can conjugate X(k) by a permutation and get

X(k)=X(k,0)00X(k,1),

where

X(k,0)=X11(k,0)X1n(k,0)Xn1(k,0)Xnn(k,0),X(k,1)=X11(k,1)X1n(k,1)Xn1(k,1)Xnn(k,1).

Let us denote the above permutation as Πk. The objective function in (4.4) is preserved if we conjugate both the C(k)’s and the X(k)’s by Πk. I.e.,

trCkXk=trΠkCkΠkTΠkXkΠkT=trCk,0Xk,0+trCk,1Xk,1,

where C(k,0) and C(k,1) are blocks corresponding to X(k,0) and X(k,1), respectively. We apply the same permutation to the X(k)’s in the constraints of (4.4) and reduce (4.4) to

minimizeXk,0,Xk,1k=0ttrCk,0Xk,0+trCk,1Xk,1subject toXk,0_0,Xk,1_0Xiik,0=Ik×k,Xiik,1=Ik+1×k+1k=0t(3)tktk!k+12trWk,0α,β,γ*Xijk,0+trWk,1α,β,γ*Xijk,10α,β,γΩtXij0,1=1Qij=𝒜EqXij1,0,Xij1,1,Qij_0, trQij=1, (5.1)

where 𝒜Eq defines the linear relationship between Qij and X(1,0),X(1,1) as

Y(ij)=[i2012010i2012]*[Xij(1,1)(1,1)0Xij(1,1)(1,2)0Xij(1,0)0Xij(1,1)(2,1)0Xij(1,1)(2,2)][i2012010i2012],
Qij=14[1+Y11(ij)+Y22(ij)+Y33(ij)00(Y31(ij)Y13(ij))01Y11(ij)Y22(ij)+Y33(ij)(Y31(ij)+Y13(ij))00(Y31(ij)+Y13(ij))1+Y11(ij)Y22(ij)Y33(ij)0(Y31(ij)Y13(ij))001Y11(ij)+Y22(ij)Y33(ij)].

5.3.2. Rounding.

(5.1) gives us the X(k,0)’s and the X(k,1)’s. We aggregate X(1,0) and X(1,1) into the 3n×3n matrix X(1). From the X(1)’s, we build the synchronization matrix S described in [39] and apply the cryo_syncrotations function on S from the ASPIRE software package [5] to recover α1,β1,γ1,,αn,βn,γn. Note that we do not truncate the eigenvectors and eigenvalues from X(1).

6. Implementation and results for synthetic cryo-EM datasets

In this section, we give a brief history of methods used for determining orientations of cryo-EM projection images, and where we stand against the current state-of-the-art.

In 1986, Vainshtein and Goncharov developed a common-lines based method for ab-initio modeling in cryo-EM [45]. In 1987, M. Van Heel also independently discovered the same method, and coined it angular reconstitution [22]. Recall that by the Fourier slice theorem, two cryo-EM projection images (in Fourier space) must intersect along a common line. Given three cryo-EM images from different viewing directions, their common lines must uniquely determine their relative orientations up to handedness. (See Figure 6.1.) The orientations of the rest of the images are determined by common lines with the first three images. This is angular reconstitution in a nutshell.

Figure 6.1.

Figure 6.1.

Three cryo-EM images uniquely determine their orientations up to handedness. Image courtesy of Singer et al. [41].

In 1992, Farrow and Ottensmeyer expanded upon angular reconstitution by developing a method to sequentially add images via least squares [20]. One major drawback in sequentially assigning orientations of cryo-EM images is the propagation of error due to false common line detection. In 1996, Penczek, Zhu, and Frank tried to circumvent the issue via brute-force search for a global energy minimizer [35]. However, the search space is simply too big for that method to be applicable. In 2006, Mallick et al. introduced a Bayesian method in which common lines between pairs of images are determined by their common lines with different projection triplets [29]. In the method by Mallick et al., at least seven common lines must be correctly and simultaneously determined, which can be problematic. In 2010, Singer et al. lowered the requirement to that only two common lines need to be correctly and simultaneously determined [41]. In 2011, Singer and Shkolnisky built upon [41] by adding a global consistency condition among the orientations [39]. This method is called synchronization, and it is regarded as the current state-of-the-art.

Note that all avaible methods for orientation determination, including the NUG approach proposed in this paper, cannot be directly applied to raw experimental images. We will explain the factors preventing us from doing so in the next subsection. We therefore numerically validate and evaluate the method using synthetic data.

6.1. Shifts, CTF and contrast.

In comparison to the simplistic forward model (2.8), there are three major imperfections in the experimental cryo-EM datasets. First, the images are not centered, and the common lines will not correspond exactly even under their true orientations. Second, the images are subject to the contrast transfer function (CTF) of the electron microscope. A CTF, as a function of radial frequency, is shown in Figure 6.2. Making matters mathematically more challenging, a CTF is typically estimated per micrograph and each micrograph would have a different CTF. Thus, all images from the same micrograph are typically assigned the same CTF. We also say those images belong to the same defocus group. However, as shown in Table 2 of Section 6.4, shifts and CTFs are not detrimental to NUG’s performance. This brings us to the third obstacle, which does prevent us from directly applying NUG on raw experimental images. The ice layer in which the molecules are frozen is not uniform. It can be thicker/thinner where different projections are taken, and this effect is equivalent to scaling the projections by a factor γ>0. So, various projections have various contrasts. Those effects are typically mitigated in class averages. Class averages are formed by in-plane aligning and averaging raw images that are estimated to have similar viewing directions. It is possible to apply NUG on class averages instead of the original raw experimental images. However, the quality of the results then depends crucially on the specific class averaging procedure being used and does not provide much insight into the performance of NUG itself.

Figure 6.2.

Figure 6.2.

Example created by Jiang and Chiu available at http://jiang.bio.purdue.edu/software/ctf/ctfapplet.html

Table 2.

Effects of shifts, CTFs and contrast on the MSE (defined in (6.8)) of the NUG SDP cryo-EM. Shifts are sampled from 𝒰(-5,5), CTFs are drawn uniformly from 4 defocus groups and contrasts are sampled from 𝒰(0.5,1.5). We used 500 simulated projections of size 64 × 64.

SNR benchmark CTF CTF and shifts CTF and contrast
1/4 0.0285 0.0381 0.3382 0.7342
1/8 0.0333 0.1450 0.5741 > 2.0
1/16 0.0698 0.7631 1.3391 > 2.0
1/32 0.4387 1.9199 > 2.0 > 2.0
1/64 1.8587 > 2.0 > 2.0 > 2.0

6.2. ADMM implementation.

There are two parts that are particularly challenging for obtaining a numerical solution to the NUG SDP (4.4) and (5.1):

  • implementing a SDP solver that is scalable to real-world problems such as orientation estimation in cryo-EM,

  • computing the coefficient matrix for generic objective function fij’s.

For the cryo-EM problem, the NUG SDP is simply too big to iterate on solvers based on techniques like interior point methods. Interior point methods are known for their accuracy. However, their accuracy is achieved at the expense of computational complexity. In essence, interior point methods solve a system of linear equations that attempts to satisfy both primal and dual feasibilities [14]. That is, solvers based on interior point methods have to invert a matrix that contains both primal and dual variables. The number and size of the signals coupled with the number of inequality constraints in (4.4) and (5.1) make this inversion impractical. Instead, we use the alternating direction method of multipliers, or ADMM. As the name suggests, ADMM alternates between the primal and the different sets of dual variables. More importantly, the steps in ADMM are linear, with the exception of an eigendecomposition on the primal variable. In practice, the eigendecomposition on the primal variable is manageable.

In general, it is not possible to obtain a closed-form expression for the coefficient matrices C(k)’s. Sometimes, we need to employ numerical integration schemes over the group 𝒢 to find the C(k)’s.

6.2.1. The alternating direction method of multipliers.

In short, ADMM solves the augmented Lagrangian via iterative partial updating. We solve an unconstrained optimization problem over the objective variable, and enforce the constraints via dual variables. We will express the NUG SDP in a more general form, and derive ADMM updates in the more general setting. (4.4) and (5.1) are SDPs of the following form:

minimizeXC,Xsubject toX0𝒜E(X)=bE𝒜IXbI. (6.1)

We can think of C and X as the following block matrices:

C=C(0)C(1)C(t),X=X(0)X(1)X(t).

Note that going from (4.4) and (5.1) to (6.1), we used the fact that

trC(k)X(k)=C(k),X(k)

because the C(k)’s are Hermitian. 𝒜E and 𝒜I are linear operators that encapsulate the equality and inequality constraints, respectively. For example, we can think of 𝒜 as

𝒜X=AI1,XAIm,X,

where, for the constraint k=0tbktrρk*gmXij(k),AI(m) is the matrix containing bkρkgm at the position of Xij(k) and 0 everywhere else.

More concretely, 𝒜(X)=b is equivalent to Avec(X)=b, where

A=vecA(1)vecA(m)T

and vec vectorizes the matrix A(m) along the columns. The adjoint operator of 𝒜 is given by

𝒜*z=matATz,

where mat is the reverse operator of vec. Furthermore, we can verify

  • 𝒜(X),z=Avec(X),z=matATz,X=𝒜*(z),X,

  • 𝒜𝒜*z=𝒜𝒜*z=AATz,

  • 𝒜*𝒜(z)=𝒜*(𝒜(X))=matATAvec(X).

Now, we describe the ADMM solver outlined in [44] for SDP (6.1). ADMM is essentially a series of partial iterative updates based on the augmented Lagrangian. So, we will define the dual variables and write out the augmented Lagrangian. Then, we derive the updates by setting the gradient of the augmented Lagrangian, with respect to specific variables, to 0.

The dual variables corresponding to (6.1) are

X_0S_0bE𝒜E(X)=0yEbI𝒜I(X)0yI0.

The Lagrangian for (6.1) is

LX,S,yE,yI-bE,bI,yE,yI+X,S+𝒜E*yE+𝒜I*yI-C. (6.2)

The motivation for defining (6.2) is so that the primal variable X satisfies the constraints using the dual variables via

minimizeS0,yE,yIbImaximizeXLX,S,yE,yI. (6.3)

Notice that (6.3) is an unconstrained optimization problem over X. If X is not PSD, then there exists S0 such that LX,S,yE,yI=-. Thus, due to the inner maximization over X, (6.3) produces a solution satisfying X0. The equality and inequality constraints are enforced in a similar manner. The maximizing X must satisfy

XL=0S+𝒜E*yE+𝒜I*yI=C.

In fact, this gives us the dual problem to (6.1)

minimizeS,yE,yIδ_0(S)+δ0(yI)(bE,bI),(yE,yI)subject toS+𝒜E*(yE)+𝒜I*(yI)=C.

The augmented Lagrangian is the Lagrangian with the dual variables regularized by the Frobenius norm:

Lρ(X,S,yE,yI)(bE,bI),(yE,yI)+X,S+𝒜E*(yE)+𝒜I*(yI)C+ρ2S+𝒜E*(yE)+𝒜I*(yI)CFro 2. (6.4)

This regularization term is crucial for numerical convergence of the solver.

The updates for the dual variables are given by their individual optimality conditions.

  • Solving 0=SLρ for S and applying the PSD projection, we get
    S(k+1)=C-𝒜E*yE(k)-𝒜I*yI(k)-1ρX(k)0.
  • Solving 0=yELρ for yE, we get
    yE(k+1)=𝒜E𝒜E*-11ρbE-𝒜EX(k)-𝒜E𝒜I*yI(k)+S(k)-C.
    Note that for the NUG SDP, 𝒜E𝒜E*= because each equality constraint is on a single entry in X. So, this update simplifies to
    yE(k+1)=1ρbE-𝒜EX(k)-𝒜E𝒜I*yI(k)+S(k)-C.
  • Note that if we solve 0=yILρ for yI, we will have to invert the operator 𝒜I𝒜I*, and this is extremely computationally challenging. Instead, we add an additional regularization term ρ2yI-yI(k)λ-𝒜I𝒜I*2, where λ is the largest eigenvalue of 𝒜I𝒜I* and yI(k) is the element from the previous yI-update. Solving 0=yILρ+ρ2yI-yI(k)λ-𝒜I𝒜I*2 for yI and applying the the nonnegativity projection, we get
    yI(k+1)=[1ρλ(bI𝒜I(X(k)))+1λ𝒜I(C𝒜E*(yE(k))𝒜I*(yI(k))S(k))+yI(k)]0.

The update for the primal variable is simply a gradient descent given by

X(k+1)=X(k)+ρS(k)+𝒜E*yE(k)+𝒜I*yI(k)-C.

We initialize the variables as the following:

  • X(0)=I,

  • yI(0)=0,

  • S(0)=0,

  • yE(0)=12𝒜EC-𝒜I*yI(0)-S(0).

We apply the updates in the following order:

  1. S(k+1)=C-𝒜E*yE(k)-𝒜I*yI(k)-1ρX(k)0,

  2. yE(k+1/2)=𝒜EC-𝒜I*yI(k)-S(k+1),

  3. yI(k+1)=1ρλbI-𝒜IX(k)+1λ𝒜IC-𝒜E*yE(k+1/2)-𝒜I*yI(k)-S(k+1)+yI(k)0,

  4. yE(k+1)=𝒜EC-𝒜I*yI(k+1)-S(k+1),

  5. X(k+1)=X(k)+ρS(k+1)+𝒜E*yE(k+1)+𝒜I*yI(k+1)-C.

Note that we updated yE twice, which guarantees the solver’s convergence to the optimizer [44].

6.2.2. Fourier coefficients for bandlimited functions on SO(3).

The coefficient matrix in (3.12) is composed of

f^ij(k)=𝒢fij(g)ρk*(g)dg.

We obtained closed-form expressions for the registration problem (2.4) under the assumption that the noise in the underlying signal is Gaussian. For the cryo-EM problem (2.11), we require a single approximation of an integral using Gaussian quadrature. However, this is not always the case. For example, the noise in the cryo-EM projections is better modeled by the Poisson distribution [36]. We describe the numerical integration scheme in [28] that can be used to obtain the desired Fourier coefficients fˆij(k). We need to make the assumption that fij’s are bandlimited by T. The method described in [28] has a better computational complexity than straight forward numerical integration. We will define the quadrature, and then outline the evaluation over the quadrature.

For function f with bandlimit T, we have the equality

f^(k)=1(2T)2j1=02T1j2=02T1l=02T1bT(l)f(αj1,βl,γj2)ρk*(αj1,βl,γj2), (6.5)

where

bTl=2Tsinπ2l+14Tm=0T-112m+1sinπ2l+12m+14T,

and

αj1=2πj12T,βl=π(2l+1)4T,γj2=2πj22T,0j1,j2,l<2T.

Recall that

ρk(α,β,γ)m,m=eimαwm,mkβeimγ.

We can re-write the entries in (6.5) as

[f^(k)]m,m=1(2T)2l=02T1bT(l)wm,m(k)(βl)j2=02T1eimγj2j1=02T1eimαj1f(αj1,βl,γj2). (6.6)

By rearranging the terms in (6.6), it becomes obvious that we should compute fˆ(k) in the following order:

  1. for all 0j2,l<2T and -TmT, compute
    S1l,j2,m=12Tj1=02T-1e-imαj1fαj1,βl,γj2,
  2. for all 0l<2T and -Tm,mT, compute
    S2l,m,m=12Tj2=02T-1e-imγj2S1l,j2,m,
  3. for all -Tm,mT, compute
    [f^(k)]m,m=l=02T1bT(l)wm,m(k)(βl)S2(l,m,m).

The complexity to compute [fˆ(k)]m,m for all m,m and k is 𝒪T4; and the complexity of the straight-forward evaluation in (6.5) is 𝒪T6 [28].

6.2.3. Simplification for cryo-EM.

Recall that the objective function for cryo-EM does not depend on β. I.e.,

fα,γ=fα,β1,γ=fα,β2,γ.

(6.6) reduces to

[f^(k)]m,m=(l=02T1bT(l)wm,m(k)(βl))(1(2T)2j2=02T1eimγj2j1=02T1eimαj1f(αj1,γj2)). (6.7)

We compute (6.7) in the following order:

  1. compute
    BTk=l=02T-1bTlwm,mkβl,
  2. for all 0j2<2T and -TmT, compute
    S1j2,m=12Tj1=02T-1e-imαj1fαj1,γj2,
  3. for all -Tm,mT, compute
    S2m,m=12Tj2=02T-1e-imγj2S1j2,m,
  4. for all -Tm,mT, compute
    [f^(k)]m,m=BT(k)S2(m,m).

The complexity to compute [fˆ(k)]m,m for all m,m and k is 𝒪T3.

6.3. Rotation MSE and FSC resolution.

We assess the performance of orientation estimation method using the mean squared error (MSE), defined as follows:

MSE1ni=1nR^iRiFro2. (6.8)

Here Ri are the true rotations (which are known in the simulation setting) and Rˆi are the estimated rotations (note that we previously used the hat notation for the Fourier transform, but here it is used for estimators). Since the estimation is up to a global rotation and possibly handedness, the two sets of rotations are aligned prior to computing the MSE.

In addition, we will include the Fourier Shell Correlation (FSC) of the reconstructed structure against the known structure. We reconstruct the 3-dimensional structure (in Fourier space) using the estimated orientations to get ϕˆ, and compare it against the known structure ϕ (in Fourier space). For each spatial frequency rk, we calculate the FSC

FSCrkv2=rkϕ^vϕ*vv2=rkϕ^v2v2=rkϕv2.

We derive the resolution (in Angstroms) by interpolating the FSC until we reach an rk yielding FSCrk0.5. See [23] for a discussion on the FSC and the cutoff value.

Note that the rotation MSE is the most direct assessment of orientation estimation methods. After all, the stated objective of orientation estimation in cryo-EM is to estimate the orientations. In addition, the quality of the 3-dimensional reconstruction depends on several other factors such as the signal-to-noise ratio of the images, the number of images, the distribution of the viewing directions and the CTF of the images. We focus on the rotation MSE because the other factors mentioned are independent of the algorithm being used.

6.4. Simulated data.

With the ASPIRE software package in [5], we generate a set of 100 and a set of 500 projection images of size 129 × 129 from the Plasmodium falciparum 80S ribosome (see https://www.ebi.ac.uk/pdbe/emdb/empiar/entry/10028/.) We add Gaussian noise to the simulated projections, and then apply a circular mask to zero out the noise outside of the radius. Table 1 shows a comparison of NUG and the synchronization method for different noise levels. In particular, the estimation error of NUG at SNR=1/64 with 500 images is sufficiently low (MSE=0.05) to result in a meaningful 3-D ab-initio model (with ¡ 30 Angstrom resolution). Numerical experiments were conducted on a cluster of Intel(R) Xeon(R) CPU E7–8880v3 @ 2.30GHz totaling 144 cores and 792GB of RAM. The synchronization method is roughly two orders of magnitude faster than NUG, but NUG gives more accurate estimates at low SNR.

Table 1.

The resolutions shown are in Angstroms. At high SNR, synchronization is accurate to more decimal places. This is not surprising since we have compromised on various truncations and discretizations for computational speed, etc. However, as the noise increases, we see NUG outperforming synchronization.

100 images SNR NUG MSE sync. MSE NUG res sync res
1/1 0.0153 1.2759e-04 24.6 20.8
1/2 0.0155 1.3593e-04 21.5 20.3
1/4 0.0174 3.6615e-04 27.0 22.4
1/8 0.0192 0.0037 28.7 25.8
1/16 0.0227 0.0300 29.8 30.9
1/32 0.0298 0.1572 35.6 45.8
1/64 0.1559 2.7818 45.3 174.1
1/128 2.1239 4.1492 97.9 175.3
500 images SNR NUG MSE sync. MSE NUG res sync res
1/1 0.0125 4.1412e-05 18.1 14.6
1/2 0.0130 5.1825e-05 20.8 18.2
1/4 0.0134 1.5268e-04 21.7 17.0
1/8 0.0143 0.0018 16.4 18.1
1/16 0.0175 0.0189 20.8 17.7
1/32 0.0195 0.1559 24.6 30.7
1/64 0.0460 2.2496 29.3 71.1
1/128 1.6060 3.1661 64.2 107.6

To put the numbers in Table 1 into perspective, a MSE of 0.1, for example, can be regarded as small enough in one instance to lead to a good reconstruction, but too large in another. The 3-dimensional reconstruction from 100 images at low SNR (e.g., SNR=1/128) looks bad even if one uses the true orientations (i.e., MSE=0). On the other hand, the 3-dimensional reconstruction from 500 images at moderate SNR (e.g., SNR=1/32) would look very decent at MSE=0.1. The latter case is shown in Figure 7 in [39], where SNR=1/32 and the MSE is slightly above 0.1. The resolution measure quantifies the quality of the 3-D ab-initio model. An idea about the quality of 3-dimensional reconstruction with respect to the MSE can also be obtained from Table 3 and Figure 8 in [42]. We would like to point out that for the same computational cost, synchronization can process more images than NUG, and therefore may yield better 3-dimensional reconstructions. However, beyond the theoretical importance, NUG still might be useful in situations in which the number of images is limited (e.g., when there are only a handful number of class averages or in electron tomography).

We also illustrate the effects of contrast, CTF and shifts on the performance of NUG in the numerical experiment reported in Table 2. In this numerical experiment, phase flipping was applied to correct for the phases of the CTF but not their magnitudes. As for shifts, we simply ignore it in the NUG SDP. In the future, we can expand (4.4) to include shifts to improve upon our estimates. Please see Section 6.1 for a discussion on the effects of contrast, CTF and shifts.

The distribution for the contrast was based on the following numerical experiment. We used the publicly available experimental dataset EMPIAR-10028 found here: https://www.ebi.ac.uk/pdbe/emdb/empiar/entry/10028/. Using the known 3-dimensional structure of the molecule, we construct 10,000 simulated clean projections of size 64 × 64 at different orientations sampled uniformly over SO(3). For each image Iˆ in the experimental dataset, we solve for γ in

minimizeγ,IiγI^IiFro 2,

where Ii’s are simulated projections generated from the known structure. Figure 6.3 shows the empirical distribution of γ.

Figure 6.3.

Figure 6.3.

Distribution of contrast in the experimental dataset EMPIAR-10028.

7. Summary

The NUG problem consists in the minimization of the sum of pairwise cost functions defined over the ratio between group elements, for arbitrary compact groups. This corresponds to the simultaneous multi-alignment of many datapoints (e.g. signals, images, or molecule densities) with respect to transformations given by an action of the corresponding group. We presented a methodology to solve this problem involving a relaxation of the problem to an SDP and an implementation of an ADMM algorithm to solve the resulting SDP.

In this paper we focused mainly in the context of alignment over SO(2) and SO(3). A notable example is that of finding a consistent set of pairwise rotations of many different functions defined on a sphere which globally minimize the disagreement between pairs of functions.

The NUG problem arises in cryo-EM, where the task is to recover the relative orientations of many noisy 2-dimensional projections images of a 3-dimensional object, each obtained from a different unknown viewing direction. Once good alignments are estimated, the 3-dimensional object can be reconstructed using standard tomography algorithms. In this paper we formulate the problem of image alignment in cryo-EM as an instance of NUG, and demonstrate the applicability to simulated datasets.

The computational and numerical considerations require truncations of the both the cost functions and group representations; a general approach is proposed as part of the formulation as an SDP, and additional methods particularly developed for the case of SO(3), and to the special properties of the cryo-EM problem, are presented.

It is noteworthy that, compared to previous work on related alignment problems, the formulation of the problem as an SDP can provide a certificate for optimality in some cases. Specifically, whenever the solution of the SDP also satisfies the nonconvex constraints that have been relaxed, it serves as a certificate that optimality has been achieved. In numerical work not reported in this paper we have numerically observed optimality of NUG SDP for some instances of MRA, but not for estimation of orientations in cryo-EM. Better theoretical understanding of when NUG SDP achieves optimality is an interesting open problem. In the context of cryo-EM, like other common-line based approaches, it does not require an initial guess. These properties make NUG a candidate for future work on robust ab initio alignment which would provide an initialization for other refinement algorithms.

Acknowledgments

The authors thank the anonymous reviewers for their valuable comments that helped improve the paper.

A. Singer was partially supported by NIGMS Award Number R01GM090200, AFOSR FA9550-17-1-0291, Simons Investigator Award, the Moore Foundation Data-Driven Discovery Investigator Award, and NSF BIGDATA Award IIS-1837992. These awards also partially supported A. S. Bandeira, Y. Chen, and R. R. Lederman at Princeton University.

Part of this manuscript was prepared while A. S. Bandeira was with the Courant Institute of Mathematical Sciences and the Center for Data Science at New York University, and partially supported by National Science Foundation Grants DMS-1317308, DMS-1712730, DMS1719545, and by a grant from the Sloan Foundation.

The authors performed most of the research at Princeton University.

Footnotes

1

Similarly to SO(2), it is possible that the non-negativity constraint may be replaced by an SDP or sums-of-squares constraint. [38]

Contributor Information

Afonso S. Bandeira, Department of Mathematics

ETH Zürich, Ramistrasse 101, 8092 Zürich, Switzerland.

Yutong Chen, Systems Trading, Tudor Investment Corporation, New York, NY USA.

Roy R. Lederman, Department of Statistics and Data Science, Yale University, 24 Hillhouse Avenue, New Haven, CT 06511 USA

Amit Singer, Department of Mathematics, Program in Applied and Computational Mathematics, Princeton University, Fine Hall, Washington Road, Princeton, NJ 08544 USA.

References

  • [1].Abbe E, Bandeira AS, Bracher A, and Singer A. Decoding binary node labels from censored edge measurements: Phase transition and efficient recovery. Network Science and Engineering, IEEE Transactions on, 1(1):10–22, Jan 2014. [Google Scholar]
  • [2].Abbe E, Bandeira AS, and Hall G. Exact recovery in the stochastic block model. IEEE Transactions on Information Theory, 62(1):471–487, 2016. [Google Scholar]
  • [3].Alon N and Naor A. Approximating the cut-norm via Grothendieck’s inequality. In Proc. of the 36 th ACM STOC, pages 72–80. ACM Press, 2004. [Google Scholar]
  • [4].Askey R. Orthogonal Polynomials and Special Functions. SIAM, 4 edition, 1975. [Google Scholar]
  • [5].ASPIRE - algorithms for single particle reconstruction software package. http://spr.math.princeton.edu/.
  • [6].Bai XC, McMullan G, and Scheres SH. How cryo-EM is revolutionizing structural biology. Trends in Biochemical Sciences, 40(1):49–57, 2015. [DOI] [PubMed] [Google Scholar]
  • [7].Bandeira AS. Random Laplacian matrices and convex relaxations. Foundations of Computational Mathematics, 18(2):345–379, 2018. [Google Scholar]
  • [8].Bandeira AS, Charikar M, Singer A, and Zhu A. Multireference alignment using semidefinite programming. 5th Innovations in Theoretical Computer Science (ITCS 2014), 2014. [Google Scholar]
  • [9].Bandeira AS, Kennedy C, and Singer A. Approximating the little Grothendieck problem over the orthogonal and unitary groups. Math. Program, 160:433–475, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Bandeira AS, Singer A, and Spielman DA. A Cheeger inequality for the graph connection Laplacian. SIAM J. Matrix Anal. Appl, 34(4):1611–1630, 2013. [Google Scholar]
  • [11].Bannai E and Bannai E. A survey on spherical designs and algebraic combinatorics on spheres. European Journal of Combinatorics, 30(6):1392–1425, 2009. [Google Scholar]
  • [12].Bhamre T, Zhao Z, and Singer A. Mahalanobis distance for class averaging of cryo-EM images. 2017 IEEE 14 th International Symposium on Biomedical Imaging, pages 654–658, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Lederman RR and Singer A. A representation theory perspective on simultaneous alignment and classification. Applied and Computational Harmonic Analysis, 2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Boyd S and Vandenberghe L. Convex Optimization. Cambridge University Press, 2004. [Google Scholar]
  • [15].Charikar M, Makarychev K, and Makarychev Y. Near-optimal algorithms for unique Games. Proceedings of the 38th ACM Symposium on Theory of Computing, 2006. [Google Scholar]
  • [16].Chen Y, Huang Q-X, and Guibas L. Near-optimal joint object matching via convex relaxation. Proceedings of the 31st International Conference on Machine Learning, 32(2):100–108, 2014. [Google Scholar]
  • [17].Coifman RR and Weiss G. Representations of compact groups and spherical harmonics. L’Enseignement Mathematique, Vol. 14, 1968. [Google Scholar]
  • [18].Cucuringu M. Synchronization over Z2 and community detection in signed multiplex networks with constraints. Journal of Complex Networks, 2015. [Google Scholar]
  • [19].Dumitrescu B. Positive Trigonometric Polynomials and Signal Processing Applications. Springer, 2007. [Google Scholar]
  • [20].Farrow NA and Ottensmeyer EP. A posteriori determination of relative projection directions of arbitrarily oriented macromolecules. Journal of the Optical Society of America A, 9(10):1749–1760, 1992. [Google Scholar]
  • [21].Goemans MX and Williamson DP. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. Journal of the Association for Computing and Machinery, 42(6):1115–1145, 1995. [Google Scholar]
  • [22].Van Heel M. Angular reconstitution: a posteriori assignment of projection directions for 3D reconstruction. Ultramicroscopy, 21:111–123, 1987. [DOI] [PubMed] [Google Scholar]
  • [23].Van Heel M and Schatz M. Fourier shell correlation threshold criteria. Journal of Structural Biology, 151:250–262, 2005. [DOI] [PubMed] [Google Scholar]
  • [24].Huang Q-X and Guibas L. Consistent shape maps via semidefinite programming. Computer Graphics Forum, 32(5):177–186, 2013. [Google Scholar]
  • [25].Kam Z. The reconstruction of structure from electron micrographs of randomly oriented particles. Journal of Theoretical Biology, 82(1):15–39, 1980. [DOI] [PubMed] [Google Scholar]
  • [26].Khot S. On the power of unique 2-prover 1-round games. Thiry-fourth annual ACM symposium on Theory of computing, 2002. [Google Scholar]
  • [27].Khot S. On the unique Games conjecture (invited survey). In Proceedings of the 2010 IEEE 25th Annual Conference on Computational Complexity, CCC ‘10, pages 99–121, Washington, DC, USA, 2010. IEEE Computer Society. [Google Scholar]
  • [28].Kostlee PJ and Rockmore DN. FFTs on the rotation group. Journal of Fourier Analysis and Applications, 14:145–179, 2008. [Google Scholar]
  • [29].Mallick SP, Agarwal S, Kriegman DJ, Belongie SJ, Carragher B, and Potter CS. Structure and view estimation for tomographic reconstruction: A Bayesian approach. EEE Conference on Computer Vision and Pattern Recognition, 2:2253–2260, 2006. [Google Scholar]
  • [30].Natterer F. The Mathematics of Computerized Tomography. Classics in Applied Mathematics, SIAM, 2001. [Google Scholar]
  • [31].Naor A, Regev O, and Vidick T. Efficient rounding for the noncommutative Grothendieck inequality. In Proceedings of the 45th annual ACM symposium on Symposium on theory of computing, STOC ‘13, pages 71–80, New York, NY, USA, 2013. ACM. [Google Scholar]
  • [32].Nesterov Y. Semidefinite relaxation and nonconvex quadratic optimization. Optimization Methods and Software, 9(1–3):141–160, 1998. [Google Scholar]
  • [33].Parrilo PA. Structured semidefinite programs and semialgebraic geometry methods in robustness and optimization. Thesis, California Institute of Technology, 2000. [Google Scholar]
  • [34].Parrilo PA Blekherman G and Thomas RR. Semidefinite Optimization and Convex Algebraic Geometry. MOS-SIAM Series on Optimization, 2013. [Google Scholar]
  • [35].Penczek PA, Zhu J, and Frank J. A common-lines based method for determining orientations for n > 3 particle projections simultaneously. Ultramicroscropy, 63:205–218, 1996. [DOI] [PubMed] [Google Scholar]
  • [36].Prucnal PR and Saleh BEA. Transformation of image-signal-dependent noise into image-signal-independent noise. Optics Letters, 6:316–318, 1981. [DOI] [PubMed] [Google Scholar]
  • [37].Rodrigues O. Des lois géometriques qui regissent les déplacements d’ un systéme solide dans l’ espace, et de la variation des coordonnées provenant de ces déplacement considérées indépendent des causes qui peuvent les produire. J. Math. Pures Appl, 5:380–440, 1840. [Google Scholar]
  • [38].Saunderson J, Parrilo PA, and Willsky AS. Semidefinite descriptions of the convex hull of rotation matrices. SIAM J. Optim, 25(3):1314–1343, 2015. [Google Scholar]
  • [39].Shkolnisky Y and Singer A. Viewing direction estimation in cryo-EM using synchronization. SIAM J. Imagin Sciences, 5(3):1088–1110, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [40].Singer A. Angular synchronization by eigenvectors and semidefinite programming. Applied and Computational Harmonic Analysis, 30(1):20–36, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [41].Singer A, Coifman RR, Sigworth FJ, Chester DW, and Shkolnisky Y. Detecting consistent common lines in cryo-EM by voting. Journal of Structural Biology, 169(3):312–322, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Singer A and Shkolnisky Y. Three-dimensional structure determination from common lines in cryo-EM by eigenvectors and semidefinite programming. SIAM J. Imaging Sciences, 4(2):543–572, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [43].Singer A, Zhao Z, Shkolnisky Y, and Hadani R. Viewing angle classification of cryo-Electron Microscopy images using eigenvectors. SIAM Journal on Imaging Sciences, 4:723–759, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [44].Sun D, Toh KC, and Yang L. A convergent 3-block semiproximal alternating direction method of multipliers for conic programming with 4-type constraints. SIAM J. Opt, 25(2):882–915, 2015. [Google Scholar]
  • [45].Vainshtein BK and Goncharov AB. Determination of the spatial orientation of arbitrarily arranged identical particles of unknown structure from their projections. 1986.
  • [46].Vanderberghe L and Boyd S. Semidefinite programming. SIAM Review, 38:49–95, 1996. [Google Scholar]
  • [47].Wang L and Sigworth FJ. Cryo-EM and single particles. Physiology, 21:13–18, 2006. [DOI] [PubMed] [Google Scholar]
  • [48].Wigner EP. Group Theory and its Application to the Quantum Mechanics of Atomic Spectra. New York: Academic Press, 1959. [Google Scholar]
  • [49].Yershova A, Jain S, LaValle SM and Mitchell JC. Generating Uniform Incremental Grids on SO(3) Using the Hopf Fibration. The International journal of robotics research, 29(7):801–812, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [50].Zhao Z, Shkolnisky Y, and Singer A. Fast steerable principal component analysis. IEEE Transactions on Computational Imaging, 2:1–12, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [51].Zhao Z and Singer A. Rotationally invariant image representation for viewing direction classification in cryo-EM. Journal of Structural Biology, 186(1):153–166, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES