Abstract
Background and Objective:
Wilson statistics describe well the power spectrum of proteins at high frequencies. Therefore, it has found several applications in structural biology, e.g., it is the basis for sharpening steps used in cryogenic electron microscopy (cryo-EM). A recent paper gave the first rigorous proof of Wilson statistics based on a formalism of Wilson’s original argument. This new analysis also leads to statistical estimates of the scattering potential of proteins that reveal a correlation between neighboring Fourier coefficients. Here we exploit these estimates to craft a novel prior that can be used for Bayesian inference of molecular structures.
Methods:
We describe the properties of the prior and the computation of its hyperparameters. We then evaluate the prior on two synthetic linear inverse problems, and compare against a popular prior in cryo-EM reconstruction at a range of SNRs.
Results:
We show that the new prior effectively suppresses noise and fills-in low SNR regions in the spectral domain. Furthermore, it improves the resolution of estimates on the problems considered for a wide range of SNR and produces Fourier Shell Correlation curves that are insensitive to masking effects.
Conclusions:
We analyze the assumptions in the model, discuss relations to other regularization strategies, and postulate on potential implications for structure determination in cryo-EM.
1. Introduction
In his seminal 1942 paper, Arthur Wilson showed that the power spectrum of molecules is approximately flat at high frequencies [23]. The core assumption Wilson made is that atoms act as if they were randomly positioned at high frequency. Wilson used that randomness assumption and the observation that atom positions map to waves in the spectral domain to argue that the cross-terms of the waves sum up incoherently, implying a flat power spectrum. Wilson statistics finds several applications in structural biology. For example, in the field of cryogenic electron microscopy (cryo-EM), it is the basis of B-factor sharpening [15], a post-processing step aimed at increasing the contrast of experimentally obtained reconstructions.
A recent paper [18] provided a rigorous derivation of Wilson statistics based on a formalism of Wilson’s argument. Using that formalism, the author derived other forms of statistics, particularly a mean and covariance estimate for the scattering potential of molecules. In the present paper, we develop a prior based on these estimates and study their use in Bayesian inference of protein structure in cryo-EM. After we state background on the use of maximum a posteriori estimation and priors in cryo-EM in Section 2.1, we describe the new prior in Section 2.2. Then, in Section 3, we evaluate the prior on synthetic linear inverse problems and discuss its properties and further applications in Section 4.
The code to reproduce numerical experiments appearing in this manuscript is publicly available at: https://github.com/ma-gilles/wilson_prior.
2. Background
2.1. Priors in cryo-EM
In standard single particle cryo-EM structure determination, the 3D scattering potential map of a molecule is determined from particle projection images . The critical computational step for reconstruction is the iterative refinement, typically formulated as maximum a posteriori estimation. In that step, the reconstructed molecule is the one that maximizes the posterior probability:
(1) |
where p(y | ϕ) is the likelihood function (the conditional probability of the images given the molecule), and p(ϕ) is the prior distribution over molecules. The prior encodes our belief about the distribution of molecules before any observation and often imposes particular properties on the inferred scattering potential (e.g., smoothness). Priors are a form of regularization: they add information to solve an ill-posed problem and avoid overfitting.
In many inverse problems, including cryo-EM reconstructions, the quality of the inferred estimate depends heavily on the choice of prior. Thus, the cryo-EM community has given much attention to crafting appropriate priors. One of the earliest software packages to use MAP estimation for 3D reconstruction in cryo-EM was RELION [16]. The initial version of RELION used a Gaussian prior distribution where each frequency is independent with zero mean and variance equal to the spherically-averaged power spectrum of the molecule at that frequency. We call this prior the diagonal prior, denoted as:
(2) |
where denotes the Fourier transform of ϕ, denotes a normal distribution with mean μ and covariance Σ, Dv denotes a diagonal matrix with diagonal v, and denotes the spherically-averaged power spectrum of ϕ. The diagonal prior has several attractive properties: it enforces smoothness on the reconstructed molecule, and it is computationally cheap thanks to the independence assumption of Fourier components.
The initial version of cryoSPARC [10], another popular software for cryo-EM reconstruction, used a spatially-independent exponential (SIE) prior:
(3) |
where Exp(μ) denotes the exponential distribution with mean μ. The advantages of this prior are that it imposes positivity on the scattering potential, and it is computationally convenient thanks to the independence of voxel values. We note that despite the statistical interpretation of priors in (Eq. 1), both the diagonal and SIE priors are chosen out of mathematical and computational convenience rather than a belief about the distribution of molecules. This may bias the reconstruction process [3]. In newer versions of RELION and cryoSPARC, both software packages use a modified version of the diagonal prior as defined above. In that prior, each Fourier frequency is modeled as independent Gaussian with variance set proportionally to the SNR as estimated by Fourier Shell Correlation (FSC) between half maps [11,17].
In recent years, implicit regularization schemes have become popular in the computational imaging community, where there is often no explicit prior function p(ϕ). Instead, an operator mimics the regularizer’s action in an iterative algorithm (e.g., the operator may act as a gradient descent step [13,14] or a proximal operator [5,20]). One example in cryo-EM, implemented in RELION, uses a denoising neural network to regularize expectation-maximization iterations [8]. Another notable example, implemented in the non-uniform refinement option of cryoSPARC, regularizes the iteration using a smoothing kernel whose parameters are set adaptively by cross-validation [12]. Implicit regularization schemes are more general than regularization using an explicit prior1 and can leverage powerful machine learning techniques that sometimes yield impressive results, e.g. [8,12] both report that the prior nearly halves the attained resolution on some test cases compared to the traditional prior. On the other hand, implicit regularization schemes lose most theoretical guarantees and statistical interpretation granted by Bayesian inference, sometimes leading to overfitting. Overfitting is particularly problematic in cryo-EM, where ground truth about the 3D structure is often unavailable, making diagnosing overfitting difficult. In the rest of this paper, we focus on building an explicit prior, similar to the ones in Eqs. (2) and (3) but derived from Wilson statistics.
2.2. The Wilson prior
The formalism used to derive Wilson statistics in [18] is the random “bag of atoms” model. In that model, a molecule consists of a collection of Natoms atoms whose positions are independently and identically distributed (i.i.d.) with probability density function that models the molecule’s shape. The scattering potential of a molecule is modeled as
(4) |
where f is the scattering potential of a single atom. For simplicity, here we assume that the atoms are identical and revisit to this assumption in Section 4. If f, g, and Natoms are fixed, Eq. (4) implicitly defines a probability distribution over molecules that can, in principle, be used as a prior for Bayesian inference. Unfortunately, there is no easy formula for the probability density function of that distribution, so it is impractical to do so. Instead, we use the first two moments of the bag of atoms model to craft a Gaussian prior that resembles the original distribution but is easier to compute. We express this prior in the Fourier domain, but an equivalent formulation in the spatial domain is described in Section A.1. The first two moments of the distribution of the Fourier transform of ϕ are derived in [18]:
(5) |
(6) |
where , denote the Fourier transforms of f and g respectively. We thus define the Wilson prior as:
(7) |
Due to the non-Gaussianity of the bag of atoms model, the Wilson prior and the bag of atoms model are not the same. In particular, negative values in the spatial domain are possible in the Wilson prior but not in the bag of atoms model (assuming f(x) ≥ 0). Nevertheless, the Wilson prior is the maximum entropy distribution, which matches the first and second-order statistics derived from the bag of atoms model and allows us to derive a fast algorithm. Fig. 1 illustrates the differences between the bag of atoms model and three priors: Wilson, diagonal, and SIE, as defined in Eqs. (7), (2) and (3).
The diagonal and SIE priors represent two extremes: in the diagonal prior, frequencies in the spectral domain are independent, and in the SIE prior, voxels in the spatial domain are independent. The Wilson prior is a middle ground: the decay of the Fourier-transformed shape function dictates the correlation between two frequencies, and the decay of the function f(x) dictates the correlation between two positions. This latter fact is most evident from the Wilson statistics expressed in the spatial domain (see Section A.1 in the Appendix). In Fig. 1, observe that projection images generated by the diagonal prior have a similar texture as the ones generated by the Wilson prior, but they do not capture the shape of the molecule. In Section Appendix A, we explain this phenomenon by showing that the diagonal and Wilson priors are equal in the limit on a uniform shape function on the entire domain. As a result, we interpret the Wilson prior as a generalization of the diagonal prior beyond the case of a uniform shape function on the entire domain.
If no information about the molecule’s shape is available, an uninformative prior reflecting equal probability of the position of atoms at any point in the domain is appropriate. Thus, the diagonal prior is the logical choice. However, during the iterative refinement of 3D reconstruction, we can often observe the low frequency of the shape function at early iterations. Remarkably, these low frequencies capture most of the correlation of high frequencies. This is a consequence of a result in [18]: under mild assumptions on the shape function, its Fourier transform decays quadratically with frequency . At high frequency, the decay implies that therefore the covariance of the Wilson statistics is:
Since decays slowly in the spectral domain, the quadratic decay of the shape function dictates the correlation of neighboring frequencies. Therefore, Fourier coefficients at high frequencies are correlated to their neighboring Fourier coefficients, but that correlation decays quadratically with the neighborhood size. It follows that capturing for small ξ (that is, the low frequencies of the shape function) is sufficient to capture most of the correlation between high frequencies. It is this correlation we seek to exploit in developing the Wilson prior.
3. Application to inverse problems
3.1. Problem statement
We evaluate the Wilson prior on linear inverse problems; that is, by inferring the 3D scattering potential of the molecule ϕ from measurements y, generated by:
(8) |
where A is the measurement matrix, and is additive Gaussian white noise. In particular, we study the denoising problem where A is the identity and the deconvolution problem where A is diagonal in the Fourier domain. Both problems are significantly simpler than 3D reconstruction of cryo-EM structures, but the expectation step of expectation-maximization takes the form of Eq. (8); where the diagonal matrix A also accounts for marginalizing over the latent variables.
We estimate ϕ from y with MAP estimation as in Eq. (1). When the noise and the prior are Gaussian distributions (as in the case for Wilson and diagonal priors), the optimization problem is rewritten as a least-squares problem:
(9) |
where μ, Σ are the prior mean and prior covariance. The solution of Eq. (9) is the result of the Wiener filter:
(10) |
Plugging in different prior means and covariances result in different MAP estimates, which we refer to as MAP-diagonal and MAP-Wilson. Both the diagonal and the Wilson priors have hyperparameters that depend on the true molecule: the spherically-averaged power spectrum in the case of the diagonal prior, the shape function g, atomic scattering function f, and the number of atoms in the case of Wilson prior. In cryo-EM applications, the spherically-averaged power spectrum can often be well approximated from the data before the reconstruction is performed [19]. However, the shape of the molecule is not typically known a priori. We address this in the following section.
3.2. Estimating parameters of the Wilson prior
The Wilson prior depends on three quantities: the number of atoms Natoms, the atom scattering function f(x) and the shape function g(x). To fix Natoms and f(x), we assume that the atomic composition of the target molecule is known. We set Natoms to the ground truth number of non-hydrogen atoms and set f equal to the weighted average of the scattering functions of atoms in the molecule:
where a indexes the atom type, pa denotes the proportion of atom a in the molecule, and fa the scattering potential of atom a, each modeled by a sum of 5 Gaussians as in [9].
To approximate the shape function g, we assume that we have a (possibly low resolution) estimate of the molecule ϕguess. We get a formula for g by approximating the mean of the Wilson statistics as ϕguess:
Solving for g(x), we get: , i.e., g(x) is proportional to our guess image deconvolved by the atom scattering function. If ϕguess were precisely the true molecule scattering function and all atom scatterings were equal to f(x), then the estimated g(x) would be a sum of delta function at the true atomic positions. To account for noise and modeling mismatch, we propose to relax this approximation by convolving this quantity with a Gaussian kernel . Finally, since g(x) is a probability distribution of the position of atoms, it must lie in the probability simplex . Numerically, this condition ensures that the covariance matrix in Eq. (6) is positive semi-definite. To enforce this constraint, we orthogonally project our guess onto Δ (denoted ΠΔ). In summary, we estimate the shape function g from a guess of the molecule ϕguess by:
(11) |
In all numerical experiments, we take ϕguess to be the MAP-diagonal estimate and set the kernel width ν = 1Å. We discuss alternatives to estimate g(x) in Section 4.
We mention that convolving an empirical distribution by a kernel is known as kernel density estimation; a non-parametric method typically used to estimate a random variable’s probability density function based on a finite data sample (e.g., [6]). Here, we use it in a different framework since we never directly observe samples of atomic locations. The effect of kernel density estimation is to smooth the distribution, which is interpreted as regularizing the estimated distribution. This strategy also recalls the regularization scheme of [12], where a non-stationary Butterworth kernel is used to regularize in cryo-EM reconstruction. A significant difference is that we use the kernel to set a hyperparameter of the prior, whereas [12] uses the kernel directly to regularize the molecule.
3.3. Evaluation of the Wilson prior on synthetic data
We construct a ground truth scattering potential ϕ from the reported atomic coordinates of the spike protein (PDB 6VXX) by evaluating the following sum on a regular 3D frequency grid using the NUFFT [1,2]:
(12) |
where a is the atom type, and is the atomic scattering model reported in [9]. The final ground truth is obtained by inverse discrete Fourier transform and masking using the ground truth mask reported below, which suppresses oscillations caused by truncation of the spectrum.
We report the Fourier Shell Correlation (FSC) between the ground truth and reconstructed molecules, both with and without masks. Masked FSC is the standard method to estimate resolution in cryo-EM, but using an overly tight mask will inflate the FSC, resulting in an over-optimistic resolution estimate. On the other hand, using no masks often underestimates the map’s resolution, as the noise in the background dominates the FSC at higher frequencies. We use the resolution criterion “masked FSC equal to 0.5” since we compare our reconstruction to the ground truth [15]. We use a ground truth mask generated by the software package EMDA [22] by placing a sphere of radius 3Å at each atomic location, followed by dilation and convolution with a Gaussian.
In Fig. 2, we compare the result of the denoising Wiener filter using the diagonal and Wilson prior at different SNRs. We define where N = 3013 is the number of voxels. Following the implementation in [16], we scale the covariance matrix of the diagonal prior (but not the Wilson prior) by a constant T = 4. This scaling factor improves the reconstruction of the MAP-diagonal estimate, as was also empirically observed in [16]. The first noteworthy result is that the FSC of the MAP-Wilson estimate is insensitive to the mask compared to the MAP-diagonal estimate. In Section A.1 in the appendix, we show that if we take the atomic scattering function to be a delta function (which is approximately true for low-resolution problems), the support of the shape function contains the support of the MAP-Wilson estimate. Hence, estimating g is analogous to calculating a mask, and the MAP-Wilson estimate has a masking effect.
Within the mask, the MAP-Wilson’s resolution is significantly higher than the MAP-diagonal: 2.6Å vs 3.1Å SNR = 10−2, and 8.1Å vs 10.1Å at SNR = 10−3. We interpret that the diagonal prior over-smoothes the protein, whereas the Wilson prior exploits the correlation of neighboring Fourier components to suppress more noise while retaining higher frequency information. Note that at SNR = 10−3, the resolutions are very low for both estimates (10.1Å for diagonal vs. 8.1Å for Wilson), but a notable difference is the MAP-Wilson estimate is visually not smooth, unlike the MAP-diagonal estimate. In this case, the smoothing of the diagonal prior might be a desirable property. We come back to this point in Section 4.
In Fig. 3, we consider the problem of deconvolution with the contrast transfer function (CTF):
with λ = 2.51 pm, d = 1.5 μm, Cs = 2.0 μm, α = 0.1, and B = 80Å2. Note that the zero crossings of the CTF annihilate some of the measured Fourier coefficients; thus, the SNR is small around these frequencies. This produces the visible oscillations in the unmasked FSC of the MAP-diagonal estimate. In contrast, the oscillations are not present in the MAP-Wilson case. Therefore, we interpret that the Wilson prior uses the correlation between neighboring frequencies to fill zero-crossing gaps. The ability to fill low-SNR gaps may be helpful in cryo-EM applications as SNR is non-uniform in Fourier space due to the CTF and the non-uniform distribution of molecule rotations.
Similar to the denoising case, the FSC of the MAP-Wilson is insensitive to masking, and the resolution of the MAP-Wilson estimate is higher than the MAP-diagonal across all SNRs: 2.2Å vs 2.7Å at SNR = 10, 2.8Å vs 3.2Å at SNR = 1 and 3.5Å vs 3.9Å at SNR = 0.1.
3.4. Computational matters
The Wiener filter in Eq. (10) involves computing the result of
(13) |
Σdiagonal and A are diagonal matrices in the Fourier basis, therefore the MAP-diagonal estimate can be computed by elementwise multiplication in operations, where N is the number of voxels. However, ΣWilson is dense, so the same strategy does not apply. Instead, we solve the linear system in Eq. (13) with the conjugate-gradient algorithm (CG) [7]. The computational bottleneck of CG is the matrix-vector products with the matrix AΣA* + σ2I. Thanks to the special structure of ΣWilson2 these matrix-vector products can be computed in using the FFT. The computational cost of solving Eq. (13) approximately is thus where κ is the number of CG iterations. In experiments, we terminate CG when κ = 100 or we achieve a tolerance of 10−8. Finally, computation of the shape function as described in Eq. (11) costs operations with the FFT and fast projection algorithms for the probability simplex [21].
We note that although the diagonal prior is faster than the Wilson prior3, solving Eq. (13) for either prior is a negligible portion of the computation performed in 3D refinement. Indeed, the dominant cost of the maximization step of a typical expectation-maximization iteration used in 3D refinement is marginalizing over latent variables and summing over projections images to form the matrix and the right-hand side, which dwarfs the cost of the CG iteration. The other additional computation is updating the prior mean and covariance (which one could consider part of the E step in an EM algorithm). Since we do not explicitly form the matrix ΣWilson, this only requires updating the shape function g. In this paper, we computed g as in Eq. (11) at negligible computational expense, but more involved strategies such as the ones discussed in section 4 could incur a more substantial cost. We leave this question to future work.
4. Discussion
We showed a prior based on Wilson statistics outperforms a diagonal prior at solving two linear inverse problems within the MAP estimation framework. We argue that the Wilson prior encodes more information about molecules, namely, the correlation between neighboring Fourier coefficients. This correlation is used to suppress noise and fill gaps in low SNR regions of the spectral domain.
In contrast with implicit regularization, whose assumptions and convergence properties are opaque, we derived the Wilson prior from first principles. It fits squarely within the MAP framework, and we can guarantee its convergence and analyze violated assumptions. The Wilson prior is the best Gaussian approximation of the bag of atoms model, which makes the following assumptions:
Assumption 1. Atoms all have the same scattering function.
Assumption 2. Atoms positions are independently identically distributed.
Assumption 3. Atoms positions are distributed with probability density function g(x).
We made Assumption 1 for convenience here, but it is not necessary, and one could extend this framework to multiple atom types. However, numerical experiments indicate that Assumption 1 is unimportant as there is a minimal loss of accuracy in denoising real molecules compared to fake molecules modified to have a single atom type.
Assumption 2 is the core assumption made in Wilson statistics, but it is violated in practice. For example, the position of subsequent atoms is fixed in an alpha-helix. Avoiding an independence assumption is difficult when crafting a prior, but considering atom interactions might bring further improvement if feasible.
Assumption 3 can be problematic as a poor choice of g(x) will lead to bias. For example, we showed that g(x) acts as a mask in the denoising case, so carefully choosing its support is crucial. We have advocated estimating g(x) from an estimate of the molecule, but this could lead to overfitting, for the same reason that crafting a mask from an estimate may lead to overfitting [19]. On the other hand, we argued in Section 2.2 that only the lowest frequencies of g(x) need to be accurately computed for most of the prior correlation to be correctly estimated. We also showed in synthetic experiments that a simple choice of g(x) leads to significant resolution improvement with no visible overfitting. It is unclear whether this result will hold when the prior is used in an expectation-maximization algorithm. In that case, a more robust framework may be to estimate g as part of the maximum likelihood estimation. The maximum likelihood estimate of g and ϕ given measurements Y is obtained by the following constrained minimization problem:
(14) |
(15) |
Minimizing this quantity with an alternating scheme (e.g., ADMM [4]) would produce iterations similar to ones outlined above: first minimize over ϕ; which is equivalent to a Wiener filter as in Eq. (10), and then minimize over g; corresponding to a non-linear version of the projection presented in Section 3.2. Our results suggest that replacing mask estimation (a routine part of 3D reconstruction) with the estimation of the shape function would be beneficial for several reasons:
It improves the resolution of the estimates.
The corresponding FSC curves are insensitive to ground truth masks.
One can estimate the shape function within the Bayesian framework, which is typically not the case for a mask.
To be integrated within a pipeline for 3D refinement, the prior requires two generalizations: a variable noise model and an adaptive variance estimation. The former is a straightforward extension of the framework presented, e.g., the noise model used in RELION can be implemented in Eq. (10) by replacing the matrix σ2 I by a different diagonal matrix. The latter is a crucial property of the default prior used in RELION and cryoSPARC: the diagonal of the prior covariance matrix is updated at each iteration with the SNR as estimated by the FSC between half maps. This strategy is a type of implicit frequency marching [3]: the regularization decreases for the high frequencies at each iteration, gradually increasing the resolution of the estimate. The same behavior does not exist for the prior we presented here since the variance of ΣWilson at high frequencies is approximately , which is independent of the current estimate. One way to achieve the same result would be to scale the covariance matrix with the factor computed by FSC. An alternative would be to treat the refinement as a blind deconvolution problem with a convolution kernel of the form . If ν were updated using the FSC at each iteration, this step would be equivalent to B-factor correction, where Wilson statistics first found applications in cryo-EM. This may suggest that B-factor correction should be performed within each iteration rather than post-processing.
5. Conclusion
We presented a prior distribution based on a simple yet expressive generative model for molecules. Then, we gave strategies to evaluate its hyperparameters and showed that, on simple inverse problems, the prior outperforms a prior similar to one often used in practice. Finally, we discussed properties of the Wilson prior, its connections to other regularization strategies, and stated potential implications for structure determination in cryo-EM.
Acknowledgment
M.A.G. and A.S. are supported in part by AFOSR FA9550-20-1-0266, the Simons Foundation Math+X Investigator Award, NSF BIGDATA Award IIS-1837992, NSF DMS-2009753, and NIH/NIGMS 1R01GM136780-01. We thank Eric J. Verbeke for valuable discussion and help in generating figures.
Appendix A
A1. The Wilson prior in the spatial domain
We establish the representation of the covariance operator ΣWilson in the spatial domain:
(A.1) |
where * denotes convolution and 〈·, ·〉 is the standard inner product on . We use the representation in the spectral domain established in [18],
(A.2) |
Let , and denote the three-dimensional Fourier transform, then:
(A.3) |
(A.4) |
(A.5) |
(A.6) |
Eq. (A.3) follows from the unitary property of the Fourier transform , Eq. (A.4) follows from the identity Cov(Aϕ) = ACov(ϕ)A*, Eq. (A.5) follows from Eq. (A.2), and Eq. (A.6) follows from the convolution theorem. Since are arbitrary, this establishes Eq. (A.1). In quasi-matrix notation, the covariance operator is written as:
where Cf denotes the convolution operator with f, Dg is the multiplication operator with g, and the outer product is defined as (uν*)ϕ = u〈ν, ϕ〉. In the denoising problem (A = I) and with the choice of f(x) = δ(x), which implies Cf = I, the Wiener filter matrix in Eq. (10) expressed in real space is:
Note that in the case, g(x) acts as a mask for the MAP-Wilson estimate:
where . That is, the MAP-Wilson estimate factorizes into to some function w multiplied by the shape function g, which implies the support of g contains the support of .
A2. The diagonal prior as the uniform limit of the Wilson prior
Recall that the Wilson and diagonal prior are defined as follows:
The diagonal of the diagonal covariance prior is the spherically-averaged power spectrum of the molecule. Under the bag of atoms model, the expected power spectrum is (see [18] for derivation):
where we have assumed that f(x) is a spherically symmetric function. We now make the dependence on explicit with superscripts, and even though that choice is not properly defined, we proceed with . In that case, the mean and covariances of Wilson and diagonal agree for ξ ≠ 0:
and similarly for ξ1, ξ2 ≠ 0:
thus . δ is not a valid choice of as there is no well-defined probability density function g for which . Instead, we consider uniform distribution on a ball of radius where is the characteristic function of the unit ball, and let the radius grow to infinity. The Fourier transform of gR is:
and we have the desired limit property . Consequently:
That is, the diagonal and Wilson priors agree in the limit of a uniform distribution on the entire domain at non-zero frequencies.
Footnotes
Declaration of Competing Interest
The authors declare no conflict of interest.
E.g., the impossibility result in [13] describes the conditions under which an implicit denoising regularizer cannot be written as an explicit regularizer.
Σ is a diagonally scaled convolution operator plus a rank one matrix. Diagonal scaling costs operations, convolution and the matrix vector product with a rank-1 matrix is , thus the total cost is .
The convergence of CG, and therefore the runtime of the Wilson prior, varies with the condition number of the matrix (AΣA* + σ2I), which depends on the CTF and the noise level. As an example, in our implementation, the denoising Wiener filter takes 26 s using the diagonal prior and 6 min using the Wilson prior for SNR = 10−2 and a grid of size 3013, on a Macbook Pro (16 GB RAM and 2 GHz Quad-Core Intel Core i5).
References
- [1].Barnett AH, Aliasing error of the kernel in the nonuniform fast fourier transform, Appl. Comput. Harmon. Anal 51 (2021) 1–16. [Google Scholar]
- [2].Barnett AH, Magland J, af Klinteberg L, A parallel nonuniform fast Fourier transform library based on an “exponential of semicircle” kernel, SIAM J. Sci. Comput 41 (5) (2019) C479–C504. [Google Scholar]
- [3].Bendory T, Bartesaghi A, Singer A, Single-particle cryo-electron microscopy: mathematical theory, computational challenges, and opportunities, IEEE Signal Process. Mag 37 (2) (2020) 58–76. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Bertsekas DP, Constrained Optimization and Lagrange Multiplier Methods, Academic press, 2014. [Google Scholar]
- [5].Buzzard GT, Chan SH, Sreehari S, Bouman CA, Plug-and-play unplugged: optimization-free reconstruction using consensus equilibrium, SIAM J. Imaging Sci 11 (3) (2018) 2001–2020. [Google Scholar]
- [6].Gramacki A, Nonparametric Kernel Density Estimation and its Computational Aspects, Springer, 2018. [Google Scholar]
- [7].Hestenes MR, Stiefel E, et al. , Methods of Conjugate Gradients for Solving Linear Systems, vol. 49, NBS; Washington, DC, 1952. [Google Scholar]
- [8].Kimanius D, Zickert G, Nakane T, Adler J, Lunz S, Schönlieb C-B, Öktem O, Scheres SH, Exploiting prior knowledge about biological macromolecules in cryo-EM structure determination, IUCrJ 8 (1) (2021) 60–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Peng LM, Ren G, Dudarev S, Whelan M, Robust parameterization of elastic and absorptive electron atomic scattering factors, Acta Crystallogr. Sect. A 52 (2) (1996) 257–276. [Google Scholar]
- [10].Punjani A, Brubaker MA, Fleet DJ, Building proteins in a day: efficient 3D molecular structure estimation with electron cryomicroscopy, IEEE Trans. Pattern Anal. Mach. Intell 39 (4) (2016) 706–718. [DOI] [PubMed] [Google Scholar]
- [11].Punjani A, Rubinstein JL, Fleet DJ, Brubaker MA, cryoSPARC: algorithms for rapid unsupervised cryo-EM structure determination, Nat. Methods 14 (3) (2017) 290–296. [DOI] [PubMed] [Google Scholar]
- [12].Punjani A, Zhang H, Fleet DJ, Non-uniform refinement: adaptive regularization improves single-particle cryo-EM reconstruction, Nat. Methods 17 (12) (2020) 1214–1221. [DOI] [PubMed] [Google Scholar]
- [13].Reehorst ET, Schniter P, Regularization by denoising: clarifications and new interpretations, IEEE Trans. Comput. Imaging 5 (1) (2018) 52–67. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Romano Y, Elad M, Milanfar P, The little engine that could: regularization by denoising (RED), SIAM J. Imaging Sci 10 (4) (2017) 1804–1844. [Google Scholar]
- [15].Rosenthal PB, Henderson R, Optimal determination of particle orientation, absolute hand, and contrast loss in single-particle electron cryomicroscopy, J. Mol. Biol 333 (4) (2003) 721–745. [DOI] [PubMed] [Google Scholar]
- [16].Scheres SH, A Bayesian view on cryo-EM structure determination, J. Mol. Biol 415 (2) (2012) 406–418. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Scheres SH, RELION: implementation of a Bayesian approach to cryo-EM structure determination, J. Struct. Biol 180 (3) (2012) 519–530. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Singer A, Wilson statistics: derivation, generalization and applications to electron cryomicroscopy, Acta Crystallogr. Sect. A 77 (Pt 5) (2021) 472–479, doi: 10.1107/s205327332100752x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [19].Singer A, Sigworth FJ, Computational methods for single-particle electron cryomicroscopy, Annu. Rev. Biomed. Data Sci 3 (2020) 163–190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Venkatakrishnan SV, Bouman CA, Wohlberg B, Plug-and-play priors for model based reconstruction, in: 2013 IEEE Global Conference on Signal and Information Processing, IEEE, 2013, pp. 945–948. [Google Scholar]
- [21].Wang W, Carreira-Perpinán MA, Projection onto the probability simplex: an efficient algorithm with a simple proof, and an application„ 2013. arXiv preprint arXiv:1309.1541 [Google Scholar]
- [22].Warshamanage R, Yamashita K, Murshudov GN, EMDA: a python package for electron microscopy data analysis, J. Struct. Biol 214 (2021) 107826. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Wilson AJC, Determination of absolute from relative X-ray intensity data, Nature 150 (3796) (1942) 151–152. [Google Scholar]