Abstract
The ability to serially interrogate single biomolecules with femtosecond X-ray pulses from free-electron lasers has ushered in the possibility of determining the three-dimensional structure of biomolecules without crystallization. However, the complexity of imaging a sample's structure from very many of its noisy and incomplete diffraction data can be daunting. In this review, we introduce a simple analogue of this imaging workflow, use it to describe a structure reconstruction algorithm based on the expectation maximization principle, and consider the effects of extraneous noise. Such a minimal model can aid experiment and algorithm design in future studies.
Keywords: X-ray free-electron laser, coherent diffraction, single particle imaging, bioimaging
1. Introduction
A potential route to crystal-free three-dimensional structure determination, studied in [1], involves injecting very many copies of a biomolecule to serially diffract from a steady stream of intense, femtosecond pulses from X-ray free-electron lasers (XFELs) [2–6]. These femtosecond X-ray pulses serve as ultrafast shutters, because photons in these pulses can effectively outrun the Coulombic explosion [7–9] and electronic damage [10] initiated on the samples that they illuminate. This idea of ‘diffraction-before-destruction’ was shown feasible for pulses in the 5–50 fs range, for soft X-rays [11] and hard X-rays [12–14]. Further demonstrations of two-dimensional imaging using XFELs include cells [15,16], viruses [17,18] and submicrometre aerosols [19].
For three-dimensional single-particle imaging using XFELs, photon-limited, two-dimensional diffraction patterns from similar biomolecules are proposed to be algorithmically oriented and assembled into a three-dimensional diffraction volume, from which the biomolecule's three-dimensional structure is recovered via iterative phasing [20]. Among numerous reconstruction algorithms designed for this task [21–24], this paper focuses on the EMC algorithm1 [25–27], which is based on the expectation maximization principle [28]. EMC succeeds even with extremely noisy patterns and additional unmeasured parameters besides orientations [29].
Ideally, because the raw data measured for reconstruction are randomly oriented, noisy spherical sections (Ewald sphere tomograms) of a minimally damaged biomolecule's Fourier intensities, we refer to this XFEL imaging technique as serial femtosecond cryptotomography (SFC; figure 1) [29]. Although the recovery of structural information from SFC data may appear forbiddingly complex, the primary ideas behind them can be explained with a minimal model. We introduce such a minimal model, use it to explain the EMC algorithm, and study the effects of extraneous photons on structure determination.
Figure 1.

Experimental set-up for serial femtosecond cryptotomography. Many copies of a biomolecule are randomly injected to intersect a regular stream of X-ray pulses (orange). Noisy, two-dimensional diffraction patterns that sample random Ewald sphere sections of the biomolecule's three-dimensional Fourier intensities are collected on a photon detector from individual biomolecules that encounter X-ray pulses. The fuzzy red cloud (far right) shows a biomolecule destroyed shortly after diffraction. The mutual orientations of many such two-dimensional patterns are algorithmically discovered, assembled and signal-averaged in three-dimension, from which the biomolecule's electron density distribution function are reconstructed via iterative phase retrieval. (Online version in colour.)
2. A two-dimensional analogue of serial femtosecond cryptotomography
We introduce SFC by summarizing its key elements as described in [25,26]. Let W denote a biomolecule's three-dimensional diffraction intensity distribution. Ewald sphere tomograms of the biomolecule are captured on photon detectors whose pixels are labelled by the index i = 1 … Mpix. If the detector samples the Ewald sphere tomogram of the biomolecule in the reference orientation at three-dimensional reciprocal space vectors qi, these samples now occur at R(Ω) · qi for a biomolecule that suffers a rotation R−1(Ω). Because W is typically band- and resolution-limited, a sufficiently fine discretization of all possible three-dimensional rotations Ωj = 1 … Mrot can represent the full set of SFC patterns W(R(Ωj) · qi).
Because each SFC scattering data K is effectively a random, photon-limited sample of the set of all possible tomograms (W(R(Ωj) · qi) ≡ Wji), we define a Poisson noise model where the probability of observing a noisy scattering pattern K of a particular tomogram W(Ωj):
![]() |
2.1 |
In principle, the formalism in the following sections remain unchanged with alternative parametrized noise models instead (see appendix A (a)). However, inaccurate models may cause less effective reconstructions.2
We now pose a pedagogical analogue that share key features with SFC. Recall that the goal in SFC is to recover the true three-dimensional intensities
from a large number of noisy, unoriented data {K}. First, we limit all data orientations to one of four that are related by π/2 rotations about a common axis. Second, instead of a set of three-dimensional diffraction intensities,
is a two-dimensional image contrast (figure 2) comprising pixels whose positions on the unrotated image are xi instead of qi. Furthermore, we project the histogram of pixel values
onto an exponential distribution ([30] describes histogram projections), to mimic that from diffraction intensities of uncorrelated scatterers3 [25]. Finally, the set of data K are still drawn from the set W(R(Ωj=0
…
3) · xi), which are two-dimensional images. We refer to these analogous K as ‘tomograms’ to be consistent with three-dimensional SFC.
Figure 2.
Schematic of the EMC reconstruction algorithm, when applied to the two-dimensional image reconstruction analogue of SFC. The iterative reconstruction of the solution contrast (top right: projection of an icosahedron with three internal spheres, with the same contrast histogram expected in SFC diffraction images) from a random initial model (top left) is depicted in the top row. The steps of each EMC iteration label the red (thin) arrows: expand the model into its four orientational ‘tomograms’; probabilistically cluster the noisy, unoriented data into these ‘tomograms’ to maximize their individual likelihood of observing the data, temporarily ignoring ‘tomogram–tomogram’ relationships; compress the ‘tomograms’ into a single updated model. Although the noisy, unoriented input datasets {K} (in bottom row, non-zero photon counts mark the otherwise empty images) describe the true image contrast in this two-dimensional analogue, similarly incomplete datasets describe a biomolecule's band-limited three-dimensional Fourier intensities in three-dimensional SFC. (Online version in colour.)
Both three-dimensional SFC and its two-dimensional image reconstruction analogue operate on unoriented data that obey the same noise model in (2.1). In both cases, we wish to identify a single model W that is most likely to result in a stream of noisy data {K} that obey (2.1). The primary difference between them is that the two-dimensional image reconstruction problem contains fewer spatial degrees of freedom.4 The datasets in both three-dimensional SFC and two-dimensional image reconstruction are redundant: when correctly oriented, any pair of {K1, K2} in SFC intersects along a common arc in the target object's three-dimensional diffraction volume [1], and in the latter pixels on an image data, K can be mapped one-to-one onto pixels on another image data via an in-plane rotation.
Although this two-dimensional image reconstruction completes when the ‘tomograms’ are assembled, a further step of iterative phase retrieval is needed for structure determination in three-dimensional SFC [26]. Although phase retrieval [31] is separated here from orientation discovery to explicitly study the latter, the interplay between the two was explored in [25].
3. The expand, maximize, compress algorithm
Reconstructing an object's intensities in SFC can be related to cluster analysis based on distribution models, where one tries to probabilistically cluster many scattering data {K} into tomograms that are consistent with the data they comprise (2.1). Only when these clusters are correctly assembled are the object's true Fourier intensities recovered.
What is notable in SFC is that these clusters are also related by rotations. Enforcing these orientational relationships between clusters allows data sharing between clusters, beneficial when the datasets {K}, and hence the clusters they form, are exceptionally noisy. Without such relationships, once the photons in noisy images have identified with an orientation cluster, they are no longer available for signal-averaging along known intersections with other orientation clusters. In principle, such data sharing can be extended to other known relationships between clusters.
Ab initio reconstruction is performed iteratively, starting from a guess model comprising a random set of numbers. We expand this guess model W into known tomograms Wj, use expectation-maximization [28] to ‘soft’-cluster {K} into these tomograms and compress these ‘maximized’ tomograms via their mutual orientation relationships into a single more compatible new model W′. In three-dimensional SFC, one can further impose that W are the band-limited diffraction intensities of a real-valued electron density distribution [25]. The output of the final compression step is now ready for another round of EMC. This is iterated until W numerically converges (figure 2).
The definiteness in the orientation Ω of each image K is replaced by a probability distribution over the orientations P(Ω|K, W) in the EMC. As the guess model W is iteratively improved, datasets {K} gradually maximize their likelihood given the noise model (2.1) by adjusting their probability masses between orientation clusters. When reconstructions succeed, P(Ω|K, W) for each data peaks around a small set of orientations [32], and approaches a delta function if K were noiseless.
Figure 2 shows a numerical simulation of a successful reconstruction with our two-dimensional analogue. A two-dimensional sample contrast, plus versions of itself rotated by π/2 were prepared. A very large, random selection of these images were Poisson-noise sampled and their orientations were forgotten. EMC was made to reconstruct the original sample contrast given only these noisy, unoriented datasets. The average data image in the bottom row of figure 2 had approximately 10 photons. A total of 2 × 105 datasets were used for the reconstruction. Specific implementation of the reconstruction is elaborated in appendix A (a), whereas further details on three-dimensional SFC can be found in [26,30]. The computer codes used to reconstruct noisy images such as those in figure 2 are publicly available.5
Reconstructions from different random sets of initial models will naturally converge in the neighbourhood of the true source image contrast if the data tomograms over-constrain the model. When reconstructions from random initializations yield vastly different reconstructions, over-fitting of model parameters may have occurred and assumptions for the datasets used in reconstruction should be revisited.
Regardless of the reconstruction algorithm used, a mutual information criteria can identify when reconstructions attempts on the true model
become ambiguous given the average number of photons N per data tomogram [25,26]. Measurements {K} can be thought of as the output of a noisy channel through which encoded information about the model
is sent. A reconstruction algorithm's task is to ‘decode’ W using these measurements {K}.
This mutual information criterion is pessimistic, because it evaluates the success based on the average data K, neglecting the fact that it could fortuitously contain a subset of ‘bright’ data Kbright sufficient for reconstructions to succeed. However, this optimism is justified only with enough data Mdata for there to be sufficiently many Kbright [32].
4. Extraneous photons
Here, we use the two-dimensional analogue of SFC to explore these consequences of extraneous photons on structure recovery, and gain insights for a more comprehensive, but computationally expensive three-dimensional study.
Typically, SFC data will contain extraneous photons scattered from upstream optics, solvent and/or carrier gases [33]. This presents two important consequences to SFC. First, owing to the randomness of sample injection (see appendix A (b)), only a fraction of SFC diffraction patterns contain scattered photons from samples, or hits, which are contaminated with extraneous photons. These hits must be separated from patterns that only contain extraneous photons, blanks. Second, even if blanks could be completely eliminated, then the extraneous photons in hits still frustrate SFC reconstructions [27]. These consequences are particularly limiting for weakly scattering nanometre-scale biomolecules when extraneous photons dominate both hits and blanks.
(a). Separation parameter
The fundamental difficulty here is that although we measure the sum of sample photons and background photons, we seek only the former. A natural scheme to separate these two photon sources subtracts the average background photon count from their sum. Intuitively, the relative error incurred in this scheme should be related to the extent that these two distributions overlap.
We propose a separation parameter, ψ, which captures the difficulty of differentiating extraneous photons from sample photons (figure 3). Suppose b and Δb denote the total number of background and sample photons in an image, respectively. The simple ratio of the difference between their photon averages to the sum of their standard deviations6 defines their separation parameter
![]() |
4.1 |
Figure 3.

Separation parameter ψ for distinguishing background photons from sample photons. Top: Poisson distributions for background (dashed) and background + sample (solid) for three cases. Bottom: contour plot for ψ as a function of the overall background level b, and the relative difference between sample and background distributions Δb. Distributions of the three coloured dots are shown in the corresponding histograms above. (Online version in colour.)
Hits and blanks are most easily identified when
, which occurs when
, or
, or both. Appendix A (c) relates ψ to the Kullback–Leibler divergence between the distributions of hits and blanks.
(b). Simultaneous hit detection and orientation discovery
We consider the case when the structure of both the background and sample is initially unknown. This occurs when extraneous photons are primarily owing to sample injection, or if the background changes considerably in time. Ideally, if an unknown contrast can be recovered from incomplete measurements by enforcing the underlying noise model (2.1), then one should also be able to recover the background if its noise model is known.
Complications, however, arise with extraneous background. While a noise model for background plus sample can be formed (A 10), the equivalent model for only sample photons is unclear. As a consequence, a naive EMC implementation can merely recover sample contrast with background, Wb + s, instead of only sample contrast, Ws. Although background subtraction could succeed when a model background is additionally recovered, Ws = Wb + s − Wb, it is not mathematically obvious how this subtracted result can improve how we orient data, because they almost never contain exclusively sample photons. Further, when a data stream is dominated by blanks or if background photons overwhelm sample photons in hits, then EMC will erroneously cluster data into tomograms based primarily on the background features.
To study some of these consequences using the two-dimensional analogue to SFC, we consider the simplistic case where extraneous photons in blanks originate from X-ray pulses interacting with stationary experimental apparatuses and hence, unlike samples, have a fixed orientation. A two-dimensional sample contrast (true contrast in top right box of figure 2) plus versions of itself rotated by π/2 (hits), and an average background pattern were prepared7 (blank). These five images were randomly sampled with replacement, to generate a stream of random, unoriented two-dimensional images that was further Poisson-sampled to simulate photon-limited noise. The average fraction of hits in this data stream was a generous 0.4 (compare with appendix A (b)).
EMC was made to cluster the datasets into five classes (Ωj
=
0
…
4)—four orientational tomograms of Wb
+
s, and a fixed-orientation background Wb—but only enforce orientation relationships between the tomograms only in the compression step. No explicit relationship between Wb
+
s and Wb was imposed. Only after reconstruction converges is the orientationally averaged background,
, subtracted from the signal-averaged Wb
+
s reconstruction. Doing so allows nominal hit detection followed by the extraction of photons owing to the sample. More sophisticated background subtraction schemes are deferred to a later study.
From figure 4, we observe that hit detection at the expense of sample contrast is still possible when the separation parameter
, as averaged over the data {K}.
Figure 4.
Simultaneously reconstructing the average background (blank), and background + sample (hit) using EMC. In the last column we subtract the former from the latter. The average number of photons in each 50 × 50 pixel blank and hit are b and b(1 + Δ) respectively. The probability of detecting true and false hits are labelled Ptruehit and Pfalsehit respectively. The actual hit fraction was 0.4 in all instances. The number of datasets used (1 × 105, 2 × 105, 4 × 105, 8 × 105 descending with the rows) fixes the total photons for {K} per case. The separation parameter
listed is averaged over all the images. The top right box in figure 2 shows the true sample contrast, and the histograms of figure 3 illustrates the noisiness of hits and blanks typically used here.
To quantify the impact of ψ on EMC reconstructions, consider the quantity Ptruehit − Pfalsehit. This difference equals one when the probability of detecting true hits is perfect and no false hits were included in the Wb + s model. For this reason, we define this quantity as hit purity (elaborated in appendix A (d)). Remarkably, figure 5 shows a clear monotonic relationship between hit purity and the separation parameter ψ, and advises us to stay well above ψ > 1.
Figure 5.

Hit purity, Ptruehit − Pfalsehit, as a function of separation parameter averaged over the data
. Hit purities from EMC's hit detection in 1250 simulated reconstructions (each with 2 × 105 data) is shown by the grey discs. Hit purity changes most dramatically between
, and this trend is unchanged even if the number of data images increases to 2 × 106. The solid line indicates the hypothetical maximum hit purity possible with threshold-based hit detection. (Online version in colour.)
We measure EMC's effectiveness in hit detection against the maximum hit purity achievable if a total photon count threshold were used to discriminate hits from blanks (elaborated in appendix A (e)). Of course, determining such an optimal threshold is non-trivial when the structure of both blanks and hits are initially unknown. Figure 3 illustrates the difficulty involved: can one define a threshold photon count to separate the subpopulations of hits and blanks when only the combined histogram of their photon counts is measured? This is especially challenging when the separation parameter is small.
Figure 5 shows that for separation ψ > 1, EMC's hit-detection approaches the maximum hit purity achievable by threshold-based detection, even though an optimal threshold was not supplied to EMC initially. We speculate that EMC's poorer performance when ψ < 1 occurs, because background features become dominant in the hits, which then attract blanks towards the background + sample model, Wb
+
s. This is consistent with how Pfalsehit falls faster than Ptruehit when ψ decreases in figure 4. Nevertheless, when
, the blanks detected by EMC using only a smaller subset of the data can inform us of a hit-detection photon-threshold, which could then be applied more efficiently to more numerous data. These results from EMC could also provide a population of blanks necessary for an efficient two-pass hit detection based on data likelihood (see appendix A (f)).
5. Conclusion
Many probability distributions affect SFC such as the random orientations of individual samples, the Poisson statistics of the photons they scatter, the fluctuating character of the X-ray pulses [34], the random arrival of samples into the X-ray interaction region, the presence of extraneous background noise and particle heterogeneity [32]. Understanding the effects of these distributions, and knowing which ones dominate is critical to the success of SFC. Compared with an extensive and realistic study, the minimal model presented here serves as an efficient platform for testing our understanding of these effects, their mitigation strategies and possibilities for optimization. As an example, this paper studies how the EMC algorithm [26] can detect and reconstruct with randomly occurring sample signal despite extraneous photons.
Acknowledgements
We acknowledge the support from Lee Kuan Yew Postdoctoral Fellowship; Human Frontier Science Programme; AMOS programme within the Chemical Sciences, Geosciences and Biosciences Division of the Office of Basic Energy Sciences, Office of Science, US DOE. The author thanks the guidance from Veit Elser, who inspired many ideas in this paper, and Hyung Joo Park for his careful reading of the manuscript. Conversations with Anton Barty, Andrew V. Martin, Filipe R.N.C Maia, Tomas Ekeberg, Max Hantke and Gijs van der Schot were indispensable to the preparation of this paper.
Appendix A
(a). EMC algorithm in two-dimensional image reconstruction
In the EMC, the expansion step is a straightforward representation of the current model as measurable data tomograms: W → W(Ωj) = Wj. The expectation maximization step follows, where a new expanded model
replaces the current one Wj to extremize the log-likelihood
| A1 |
| A2 |
This extremization leads to the following update rule for W → W’:
![]() |
A3 |
with P(K|W(Ωj)) defined in (2.1). Although a uniform prior is used for P(Ω) in (A 3), we can estimate the actual distribution when the model iteratively converges:
| A4 |
After the pixels in each orientation representation W(R(Ωj=0 … 3) · xi) have been separately updated, they are compressed to concur on a common model image W. Care must be taken to ensure that the compressed model is properly normalized to be compatible with the datasets [26]. Two rudimentary checks are: merely expanding and compressing the model leaves it unchanged; the average total power of the tomograms is comparable to that of the data. Unlike three-dimensional SFC, pixels in our two-dimensional images rotate perfectly into that of any other orientation, which makes pixel interpolation [26] unnecessary. Hence, we simply average the images in the four orientational tomograms W(R(Ω) · xi) after they have been individually rotated back to the reference orientation:
![]() |
A5 |
Each reconstruction attempt begins with a random set of real numbers as the initial model W(0) to limit the model bias, because expectation maximization algorithms can get trapped in local log-likelihood extrema. To reduce the number of initial iterations necessary, we normalize the sum of contrast values on these random initial models W(0) to match the sum of the average data:
| A6 |
where MK is the number of noisy data images.
Reconstruction attempts in this paper are terminated when the model change ΔW between iterations becomes numerically negligible: ΔW = ||W’ − W|| < δcut-off. Of course, convergence does not imply reconstruction success.
(b). Considerations for optimal hit fractions
The probability of illuminating NF particles with an X-ray pulse would likely follow a Poisson distribution
![]() |
A7 |
Excepting fluctuation correlation methods [23], reconstruction algorithms such as the EMC typically operate on diffraction data {K} that are dominated by single-particle patterns. Translating this to a specific requirement, suppose we wish for
| A8 |
the maximum average number of particles illuminated by pulses
. At this hit rate, the number of single-particle patterns collected at a 120 Hz XFEL source will be 141 in a minute, 8469 in an hour and 101 627 in 12 h.
(c). Relationship of separation parameter ψ to Kullback–Leibler divergence
A natural measure of differences between two statistical distributions is their Kullback–Leibler divergence. For a well-behaved photon integrating detector, the probability that photons Ki are measured by its ith pixel, either from background (bi) or background + sample (bi(1 + Δi)) sources are
| A9 |
| A10 |
We choose the Kullback–Leibler divergence between these two distributions as
![]() |
A11 |
where b = ∑ibi and Δb = ∑i(Δibi). In the limit when
,
| A12 |
| A13 |
which gives us
| A14 |
Equation (A 14) shows a monotonic relationship between the separation parameter and the Kullback–Leibler divergence between two Poisson distributions. Using (A 1) and (A 11), it is straightforward to numerically verify that this monotonic relationship is valid for the critical range of 0 < Δ < 1, and 0 < b < 1.
(d). Hit purity
We define Ptruehit as the probability that a hit is correctly identified. Hence, the probability that a hit is incorrectly marked as a blank will be
| A15 |
A complementary relationship exists
| A16 |
Subtracting these two equations gives us a symmetric relationship:
| A17 |
which we denote hit purity. Because EMC does not assign each image as either a hit or a blank, these probabilities are computed here in an averaging fashion across all images {K}:
| A18 |
| A19 |
| A20 |
| A21 |
with P(Ω) defined in (A 4). Hit purities in figure 5 are evaluated only on the converged reconstructions W.
(e). Hit detection with an intensity threshold
Consider the naive hit-detection scheme where data are classified as blanks (i.e. only background photons are present), if their total photon count does not exceed a threshold μ. If the noise models for blanks and hits (i.e. both background and sample photons are present) are described by (A 9) and (A 10) respectively, then the hit purity in (A 17) for this threshold-based hit detection is
![]() |
A22 |
The following definitions apply here: b is the average number of background photons per blank; b(1 + Δ) is the average number of background plus sample photons per hit; k is the total number of photons on a data frame to be considered for hit/blank detection; μ is the threshold for discriminating blanks from hits.
For a particular combination of b and Δ, we can numerically evaluate
which maximizes the hit purity in (A 22). The separation parameter ψ can also be computed from (4.1). The resultant relationship between μmax and ψ is interpolated to form the solid curve in figure 5.
(f). Hit detection with a log-likelihood threshold
If we can identify a sufficiently large number of blanks, then a two-pass ‘variance-by-hand’ method can be adopted to select a threshold for detecting blanks, which also serves as hit detection to some degree. On the first pass through all blanks, {Kblanks}, an empirical histogram for PB in (A 9) can be computed. On a second pass through {Kblanks}, assuming the correlations between pixel readouts to be negligible, we compute the data log-likelihood for the pixels (indexed i = 1 … Mpix)
![]() |
A23 |
From the distribution of
for these blanks, one can identify the log-likelihood threshold above which we are 99% confident of identifying blanks. This also serves as a 1% significance test for background (i.e. Pfalseblank = 0.01)—cases with log-likelihood below this threshold are unlikely to contain background, which by default here are hits.
Endnotes
Abbreviated by the sequential operations in the reconstruction algorithm: expand, maximize, compress.
For example, approximating the sum of a small number of discrete events as a chi-squared distribution could increase the rate of false negatives in hypothesis testing, ch. 14.3.2 [35]. We expect a similar effect here if inaccurate noise models are used.
The joint probability in (2.1) is indifferent to whether the model W belongs to an image contrast or describes a distribution function of diffraction intensities in SFC. As long as their Ws share the same probability distribution, enforced here with a histogram projection, the effectiveness of using the unordered product in (2.1) for orientation discovery is in principle identical.
Because SFC reconstructions succeed with a sufficiently fine discrete sampling of orientations (discussed in [26]), this two-dimensional analogue merely reduces the computation complexity by reducing the number of orientations. Computation time scales quadratically with the number of orientations, Mrot: maintaining the same level of signal-averaging per orientation cluster, the number of data images needed scales linearly with Mrot; simplistic algorithms compare the number of images with the number of orientation clusters. We do not expect the number of EMC iterations needed for convergence to scale with Mrot when the data have sufficiently high signal-to-noise ratios.
See http://www.duaneloh.com, or email the author.
For the sake of simplicity, total number of photons received on hits and blanks are assumed to follow Poisson distributions, whose skewness is ignored in the separation parameter. Using variances in the place of standard deviations in the denominator of ψ incorrectly drops ψs needed dependence on b.
The ensemble of average backgrounds is spatially homogeneous, which makes the reconstruction more difficult than if background were confined to specific regions of the image that one may ignore.
References
- 1.Huldt G, Szöke A, Hajdu J. 2003. Diffraction imaging of single particles and biomolecules. J. Struct. Biol. 144, 219–227. ( 10.1016/j.jsb.2003.09.025) [DOI] [PubMed] [Google Scholar]
- 2.Madey JMJ. 1971. Stimulated emission of Bremsstrahlung in a periodic magnetic field. J. Appl. Phys. 42, 1906–1913. ( 10.1063/1.1660466) [DOI] [Google Scholar]
- 3.Ackermann W, et al. 2007. Operation of a free-electron laser from the extreme ultraviolet to the water window. Nat. Photon. 1, 336–342. ( 10.1038/nphoton.2007.76) [DOI] [Google Scholar]
- 4.Emma P, et al. 2010. First lasing and operation of an Ångstrom-wavelength free-electron laser. Nat. Photon. 4, 641–647. ( 10.1038/nphoton.2010.176) [DOI] [Google Scholar]
- 5.Ishikawa T, et al. 2012. A compact X-ray free-electron laser emitting in the sub-ångström region. Nat. Photon. 6, 540–544. ( 10.1038/nphoton.2012.141) [DOI] [Google Scholar]
- 6.Abela R, et al. 2006. XFEL: the European X-ray free-electron laser - technical design report. Number DESY 06-097. Hamburg, DESY.
- 7.Neutze R, Wouts T, Van Der Spoel D, Weckert E, Hajdu J. 2000. Potential for biomolecular imaging with femtosecond X-ray pulses. Nature 406, 752–757. ( 10.1038/35021099) [DOI] [PubMed] [Google Scholar]
- 8.Jurek Z, Faigel G, Tegze M. 2004. Dynamics in a cluster under the influence of intense femtosecond hard X-ray pulses. Eur. Phys. J. D Atomic 29, 217–229. ( 10.1140/epjd/e2004-00033-3) [DOI] [Google Scholar]
- 9.Bergh M, Huldt G, Timneanu N, Maia FRNC, Hajdu J. 2008. Feasibility of imaging living cells at subnanometer resolutions by ultrafast X-ray diffraction. Q. Rev. Biophys. 41, 181–204. ( 10.1017/S003358350800471X) [DOI] [PubMed] [Google Scholar]
- 10.Quiney HM, Nugent KA. 2011. Biomolecular imaging and electronic damage using X-ray free-electron lasers. Nat. Phys. 7, 142–146. ( 10.1038/nphys1859) [DOI] [Google Scholar]
- 11.Chapman HN, et al. 2006. Femtosecond diffractive imaging with a soft-X-ray free-electron laser. Nat. Phys. 2, 839–843. ( 10.1038/nphys461) [DOI] [Google Scholar]
- 12.Chapman HN, et al. 2011. Femtosecond X-ray protein nanocrystallography. Nature 470, 73–77. ( 10.1038/nature09750) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Boutet S, et al. 2012. High-resolution protein structure determination by serial femtosecond crystallography. Science 337, 362–364. ( 10.1126/science.1217737) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Barty A, et al. 2012. Self-terminating diffraction gates femtosecond X-ray nanocrystallography measurements. Nat. Photon. 6, 35–40. ( 10.1038/nphoton.2011.297) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Mancuso AP, et al. 2010. Coherent imaging of biological samples with femtosecond pulses at the free electron laser FLASH. New J. Phys. 12, 035003 ( 10.1088/1367-2630/12/3/035003) [DOI] [Google Scholar]
- 16.Marvin Seibert M, et al. 2010. Femtosecond diffractive imaging of biological cells. J. Phys. B Atomic Mol. Opt. Phys. 43, 194015 ( 10.1088/0953-4075/43/19/194015) [DOI] [Google Scholar]
- 17.Marvin Seibert M, et al. 2011. Single mimivirus particles intercepted and imaged with an X-ray laser. Nature 470, 78–81. ( 10.1038/nature09748) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Kassemeyer S, et al. 2012. Femtosecond free-electron laser X-ray diffraction data sets for algorithm development. Opt. Express. 20, 4149–4158. ( 10.1364/OE.20.004149) [DOI] [PubMed] [Google Scholar]
- 19.Loh ND, et al. 2012. Fractal morphology, imaging and mass spectrometry of single aerosol particles in flight. Nature 486, 513–517. ( 10.1038/nature11222) [DOI] [PubMed] [Google Scholar]
- 20.Chapman HN, et al. 2006. High-resolution ab initio three-dimensional X-ray diffraction microscopy. J. Opt. Soc. Am. A 23, 1179–1200. ( 10.1364/JOSAA.23.001179) [DOI] [PubMed] [Google Scholar]
- 21.Bortel G, Faigel G. 2007. Classification of continuous diffraction patterns: a numerical study. J. Struct. Biol. 158, 10–18. ( 10.1016/j.jsb.2006.10.018) [DOI] [PubMed] [Google Scholar]
- 22.Fung R, Shneerson V, Saldin D, Ourmazd A. 2008. Structure from fleeting illumination of faint spinning objects in flight. Nat. Phys. 5, 64–67. ( 10.1038/nphys1129) [DOI] [Google Scholar]
- 23.Saldin D, Shneerson V, Fung R, Our mazd A. 2009. Structure of isolated biomolecules obtained from ultrashort X-ray pulses: exploiting the symmetry of random orientations. J. Phys. Condensed Matter 21, 134014 ( 10.1088/0953-8984/21/13/134014) [DOI] [PubMed] [Google Scholar]
- 24.Schwander P, Fung R, Phillips GN, Jr, Ourmazd A. 2010. Mapping the conformations of biological assemblies. N. J. Phys. 12, 035007.. ( 10.1088/1367-2630/12/3/035007) [DOI] [Google Scholar]
- 25.Elser V. 2009. Noise limits on reconstructing diffraction signals from random tomographs. IEEE Trans. Inf. Theory 55, 4715–4722. ( 10.1109/TIT.2009.2027547) [DOI] [Google Scholar]
- 26.Loh ND, Elser V. 2009. Reconstruction algorithm for single-particle diffraction imaging experiments. Phys. Rev. E 80, 6705 ( 10.1103/PhysRevE.80.026705) [DOI] [PubMed] [Google Scholar]
- 27.Philipp HT, Ayyer K, Tate MW, Elser V. 2012. Solving structure with sparse, randomly-oriented X-ray data. Opt. Express. 20, 13 129–13 137. ( 10.1364/OE.20.013129) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Dempster A, Laird N, Rubin D. 1977. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. B 39, 1–38. [Google Scholar]
- 29.Loh ND, et al. 2010. Cryptotomography: reconstructing 3D Fourier intensities from randomly oriented single-shot diffraction patterns. Phys. Rev. Lett. 104, 25501 ( 10.1103/PhysRevLett.104.025501) [DOI] [PubMed] [Google Scholar]
- 30.Loh ND, Eisebitt S, Flewett S, Elser V. 2010. Recovering magnetization distributions from their noisy diffraction data. Phys. Rev. E 82, 061128 ( 10.1103/PhysRevE.82.061128) [DOI] [PubMed] [Google Scholar]
- 31.Elser V. 2003. Phase retrieval by iterated projections. J. Opt. Soc. Am. 20, 40–55. ( 10.1364/JOSAA.20.000040) [DOI] [PubMed] [Google Scholar]
- 32.Loh ND. 2012. Effects of extraneous noise in cryptotomography. Proc. SPIE 85000k, 1–15. [Google Scholar]
- 33.Kimura T, et al. 2013. Imaging live cell in microliquid enclosure by X-ray laser diffraction. Nat. Commun. 5, 1–7. ( 10.1038/ncomms4052) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Loh ND, et al. 2013. Sensing the wavefront of X-ray free-electron lasers using aerosol spheres. Opt. Express. 21, 12 385–12 394. ( 10.1364/OE.21.012385) [DOI] [PubMed] [Google Scholar]
- 35.Vetterling WT, Press WH, Teukolsky SA, Flannery BP. 2007. Numerical recipes, 3rd edn. New York, NY: Cambridge University Press. [Google Scholar]










