Abstract
Motivation
Cryo-Electron Tomography (cryo-ET) is a 3D imaging technology that enables the visualization of subcellular structures in situ at near-atomic resolution. Cellular cryo-ET images help in resolving the structures of macromolecules and determining their spatial relationship in a single cell, which has broad significance in cell and structural biology. Subtomogram classification and recognition constitute a primary step in the systematic recovery of these macromolecular structures. Supervised deep learning methods have been proven to be highly accurate and efficient for subtomogram classification, but suffer from limited applicability due to scarcity of annotated data. While generating simulated data for training supervised models is a potential solution, a sizeable difference in the image intensity distribution in generated data as compared with real experimental data will cause the trained models to perform poorly in predicting classes on real subtomograms.
Results
In this work, we present Cryo-Shift, a fully unsupervised domain adaptation and randomization framework for deep learning-based cross-domain subtomogram classification. We use unsupervised multi-adversarial domain adaption to reduce the domain shift between features of simulated and experimental data. We develop a network-driven domain randomization procedure with ‘warp’ modules to alter the simulated data and help the classifier generalize better on experimental data. We do not use any labeled experimental data to train our model, whereas some of the existing alternative approaches require labeled experimental samples for cross-domain classification. Nevertheless, Cryo-Shift outperforms the existing alternative approaches in cross-domain subtomogram classification in extensive evaluation studies demonstrated herein using both simulated and experimental data.
Availabilityand implementation
Supplementary information
Supplementary data are available at Bioinformatics online.
1 Introduction
Complex macromolecules participate in a large number of biochemical processes which helps them sustain their cellular environment and governs their cellular activities. For an accurate analysis of cellular processes enabled by the interaction of these macromolecules, an in-depth analysis of their spatial organizations and native structures within the cell is required. Such an analysis needs the observation of macromolecules in situ, which has been recently made possible with cryo-Electron Tomography (cryo-ET). Cellular cryo-ET is an imaging technology that helps to visualize 3D structures of subcellular structures from a series of 2D projections generated through an electron microscope in cryogenic temperature, where the native structure and orientation of sub-cellular components in the cell are preserved (Lučić et al., 2013).
With the advent of cryo-ET, many 3D structures in the form of tomograms can be swiftly generated for the analysis of macromolecules they contain. The difficulty, however, lies in decoding the structural information of the macromolecular complexes from these tomograms. These reconstructed 3D tomograms have complexities in the form of very low Signal-to-Nose Ratio (SNR) and missing wedge effects, making the structural recovery of these macromolecules inherently difficult. Classification of macromolecules in tomograms in such cases goes a long way toward their structural recovery from tomogram level data. To classify macromolecules, in a tomogram, we first extract subtomograms, sub-volumes of tomograms where a single macromolecule is most likely to be present. These subtomograms are then classified based on the macromolecule they contain. CNN-based methods of supervised and semi-supervised classification of subtomograms that excel in accuracy and speed were proposed by Xu et al. (2017), Che et al. (2018), Gao et al. (2020) and Liu et al. (2019). The classified subtomograms can be then aligned and averaged (Briggs, 2013) to get a higher resolution average of the corresponding structure.
While obtaining large-scale experimental data in the form of tomograms for classification is not an issue, the annotation of this data with corresponding macromolecule identifiers requires a considerable amount of time and compute. Common methods of annotation like template matching proposed by Best et al. (2007), Beck et al. (2009) and Kunz et al. (2015) are extremely time-consuming and require quality control in the form of inspection by experts, effectively making the entire process long-drawn and laborious. An apparent lack of scalability in data processing makes large-scale data of macromolecular complexes in the form of labeled subtomograms hard to obtain. To get around this issue, methods like unsupervised template-free subtomogram classification have been proposed by Xu et al. (2012), Bartesaghi et al. (2008) and Chen et al. (2014) while automatic annotation procedures using neural networks have been proposed by Chen et al. (2017). Active learning-based methods that make use of minimal labeled data have also been proposed by Du et al. (2021).
A potential solution to the scarcity of annotated data would be the development of classification algorithms that can classify subtomograms assisted by domain adaptation and randomization algorithms. Domain randomization methods work by training a classification algorithm on labeled simulated data that is easily available and randomizing this data so that the network trained on it can regularize easily to real data as proposed by Che et al. (2019). Contrary to changing (randomizing) the simulated data, domain adaptation-based methods train the deep learning-based classifier in such a way that it can classify both the simulated and the real data corresponding to the source and target domains. Domain in this context refers to data distributions that are related but different. The difference in data distributions is typically projected in the form of a domain shift in the extracted feature space. The concept of domain shift has been illustrated in Figure 1. One of the first adversarial domain adaptation methods for the classification of subtomograms was proposed by Lin et al. (2019) while a few shot method for adaptation was proposed by Yu et al. (2021). All of these methods of classification depend largely on the data simulation procedure used as it forms the base network trained on the source dataset. While some cryo-ET data simulation methods simulate isolated macromolecules, methods proposed by Liu et al. (2020) and Pei et al. (2016) can dynamically simulate complete tomograms by packing multiple macromolecular complexes with additional factors in the form of simulated noise, contrast transfer function and missing wedge effects.
Fig. 1.

Domain shift illustrated between source and target distributions
In this work, we propose the use of a network-driven domain randomization procedure in conjunction with an unsupervised domain adaptation algorithm for the classification of subtomograms. Fundamentally differing from Che et al. (2019) which focuses on randomizing hyperparameters and then performing simulations, our network-driven method Cryo-Shift of domain randomization involves gathering information from the already simulated data to add noise and distortions, allowing the model to generalize to the target domain. Furthermore, we propose the use of a multi-adversarial unsupervised domain adaptation framework inspired from Pei et al. (2018) that takes in minimal unannotated data from the target domain to bridge the domain shift between the simulated and real datasets. We follow the same data generation procedure as Che et al. (2019) in our experiments.
Our primary contributions, thus, can be summed up as:
The development of a network-driven domain randomization algorithm inspired by Zakharov et al. (2019) along with ‘warp’ modules that distort and alter the simulated data for helping the model generalize better.
The addition of an adversarial loss function with a discriminator to help the warp network provide ‘realistic’ warps.
The addition of an unsupervised domain adaptation algorithm that helps reduce the domain shift between real and simulated subtomograms, improving cross-domain classification. Being completely unsupervised, the algorithm provides an efficient solution to the problems faced in cryo-ET subtomogram annotation.
2 Methodology
2.1 Overview and notation
Our methods primarily serve to reduce the domain gap between simulated and experimental data via network-guided domain randomization and unsupervised domain adaptation. Our overall pipeline consists of a warp network W, a ‘step zero network’ M, a feature discriminator Fd and a domain discriminator D. Our data simulation module S takes in a set of hyperparameters h to simulate data from corresponding density maps () of n classes (in our analysis n = 4). As the first step of our pipeline, we train the base network M, consisting of a feature extractor Fe and a classifier C, with simulated data for predicting class labels . The feature extractor Fe denotes the convolutional blocks in M while the classifier C refers to the linear layers after Fe in M.
As the second step of the pipeline, we train the warp network W, the feature discriminator Fd, the domain discriminator D and retrain the base network M simultaneously. The warp network W is fed with simulated data that it distorts to return a warped set. The domain discriminator D and the warp network W are trained adversarially where D tries to classify the outputs of W as simulated and W tries to fool D into making wrong decisions. The feature extractor from the base network M, Fe, is fed with a concatenated dataset consisting of the output of the warp network , the simulated data and a subset of the real data DR. While the labels of the simulated data and the output from the warp network are automatically present, the labels of are not used here. Additionally, the output from the warp network is detached from the computational graph before being fed to Fe. The output from Fe is fed to the feature discriminator Fd and the classifier C. Here, Fd and Fe are trained adversarially while C is trained only on the labeled subset of the data from the feature extractor Fe. While Fd tries to classify the features from Fe into real and simulated domains, C tries to classify the features into classes accurately. An important difference between the domain discriminator D and the feature discriminator Fd is that the former works directly on subtomograms while the latter works on features. Thus, while the former tries to change the subtomogram to look more like real subtomograms, the latter makes sure that the features learned by the network are similar in nature. The overall algorithm is described in Algorithm 1. Figure 2 provides a diagrammatic representation of our method architecture.
Fig. 2.
Model overview. (a) Diagrammatic representation of warp modules and domain discriminator. (b) Diagrammatic representation of the multi-adversarial domain adaptation framework used. The Source Filter here refers to a filter that allows the classifier to back-propagate on loss calculated only on the simulated and the warped (source) data
Algorithm 1 .
Overall Algorithm
Input: hyperparamters h, Density map , Experimental Data DR
Output Trained Network M
1: procedure train(h,,DR)
2: Load M ▹ Pre-trained Step-Zero Network
3: split(DR) ▹ 1:9 split
4: for i in num_iterations do
5: ▹ Domain Randomization
6:
7:
8:
9:
10: ▹ Adaptation
11:
12:
13:
14:
15:
16:
17: end for
18: return M ▹ Trained network
19: end procedure
2.2 Step zero network
The step zero network or the base network is a simple classification network we pre-train on the simulated data . This network forms the baseline upon which further training takes place with the help of warp modules and adaptation algorithms. To establish the consistent nature of out algorithm, we make use of multiple architectures in our ablation experiments (in Section 2.7) while proceeding with a single architecture, CB3D from (Che et al., 2018), in the main experiment. A list of the model architectures used along with their corresponding parameters is provided in Table 1.
Table 1.
Networks used
| Architecture | Parameter count |
|---|---|
| CB3D (Che et al., 2018) | 72.2 M |
| DSRF3D_v2 (Che et al., 2018) | 18.2 M |
| DenseNet3D (Hara et al., 2018) | 11.2 M |
| ResNet3D (Hara et al., 2018) | 85.1 M |
2.3 Warp network
The Warp Network W is built of a set of warp modules that distort or alter the simulated data and help the base network generalize more to the experimental data during inference. Domain randomization in the context of the warp modules we use is a misnomer because while domain randomization modules deal with randomizing the source domain to help the model generalize better to any target domain, we specialize our warp network to work on a particular target domain. Although this restricts the portability of our model onto any target dataset, this helps us enhance the performance of our network on the single target dataset we train it on. Thus, while Zakharov et al. (2019) proposes making use of a gradient reversal layer (GRL) to oppose the classifier for training the randomization network, we propose the use of a discriminator with a GRL to oppose our warp network. Here, we pass the output of the warp network and randomly sampled data from the validation set in equal proportions to the discriminator that is tasked with classifying the data as real or fake. The presence of the GRL between the warp network and the discriminator makes the warp network inclined to fool the discriminator into classifying the warped data as real. The concept of GRL, inspired from Ganin and Lempitsky (2014) is that the gradients are reversed during backpropagation, thereby allowing the Discriminator and the warp network to train in opposing manners.
The warp network consists of an encoder–decoder-based architecture with the output of a single encoder E being fed to multiple decoders, each emulating one warping module. The encoder consists of three convolutional blocks with skip connections that connect to the decoder layers and are concatenated there. The input to the encoder is batch-wise 3D cryo-ET subtomogram data in the form of with b denoting the batchsize. The output from the encoder, the bottleneck Be can be expressed as . This bottleneck is split into two equal parts channel-wise and fed to the two decoders which act as warp modules. The warping modules we make use of are:
Noise Module (Wn)
Distortion Module (Wd)
2.3.1 Noise module
The noise module (Wn) adds noise to the input subtomogram In as the first randomization step. The encoder split is then passed through a simple decoder with three convolutional blocks where the skip connections from the encoder are concatenated at their respective layers. The output from the last decoder block is passed through a scaled sigmoid activation function, restricting its value to . The final output is the summed up with the input to the module In and is represented as .
2.3.2 Distortion module
The distortion module takes in two inputs, On from the noise module and a channel-wise split of the encoder output Edec. Edec is fed to two subnetworks: the first consisting of a single convolutional and linear layer while the second being a decoder with three convolutional blocks with the corresponding skip connections from the encoder concatenated. The first network provides an output in the form of a single value per sample per channel restricted to (0, 1), while the second decoder gives out a 3D dense grid . This dense-grid is then convolved with a 7 × 7 Gaussian 3D kernel having σ = 1 and multiplied in a channel-wise fashion with α where the three channels store values of x, y and z coordinates. The values in this new dense-grid gridd are then mapped with the coordinates in On and the corresponding deformed output subtomogram Od is created.
| (1) |
| (2) |
2.4 Domain discriminator
The Domain discriminator D consists of an encoder module, acting as a binary classifier, and is pre-trained on simulated data and the unlabeled subset of real data DR. The discriminator predicts if the data it is supplied with is from the real dataset or the simulated dataset. While it strives to reduce the discriminative loss and classify the domains correctly, the warp network training alongside adversarially, tries to fool it in classifying its output as real.
2.5 Domain adaptation
We propose an unsupervised domain adaptation paradigm inspired by Pei et al. (2018). This paradigm consists of n similar feature discriminators where n represents the number of classes our classifier has. These discriminators are simply binary classifiers that try to classify the features to their respective domains (simulated or real). A concatenated set consisting of output from the warp network, simulated subtomograms and real subtomogram data () in equal proportions is fed to the feature extractor Fe of base model M at each iteration. These features from Fe are fed to each of the n feature discriminators. The domain labels for the simulated and warped data are zero, while the domain label for the experimental data is set to one. A weighted gradient reversal layer (GRL) is set up between the discriminators and the feature extractor, forming an adversarial relationship between the networks during training. During the calculation of the mean discriminative loss for training the feature discriminators, a weighting system is used with the predicted class probabilities from the classifier C (detached from the computational graph), for the same features, acting as weights as demonstrated in Equation 5. The pre-trained classifier, which can already predict real subtomograms with considerable accuracy, comes in use now as the discriminative loss for a particular discriminator is multiplied with the probability of the subtomogram belonging to that particular class as predicted by the classifier.
2.6 Objective function
The step-zero network, M, is trained on the domain randomized simulated data using , a categorical cross-entropy loss as the loss function. We make use of this pre-trained model M in the next stage where we simultaneously train the base network M, the domain discriminator D, the warp network W and the feature discriminator Fd. The discriminative loss is a combination of the domain discriminative loss used to train D and W and the feature discriminative loss used to train Fd and Fe. As in the previous stage, in this stage also a task loss is used to train the step-zero network M (Fe and C). The corresponding loss function for the entire algorithm can be expressed as:
| (3) |
| (4) |
where y represents the task ground truth. td and tf represent the domain ground truth for the two discriminators.
The components of and can be represented as:
| (5) |
| (6) |
where Fi represents the individual discriminators from 1 to N with N denoting the total classes (four in our case). yj represents the one-dimensional features from Fe for the jth sample with X denoting the total samples in a batch. probj represents the class confidence of the ith class for the jth sample.
2.7 Training
We follow the methodology described above and pre-train the step-zero network as a basic classification module. We use Adam as an optimizer with a learning rate of 1e-4 and set betas to 0.9 and 0.999. We make use of a similar optimizer in the second phase of the training with the learning rate set to 1e-5 for the step zero network and the learning rates for the warp network, domain discriminator and the feature discriminator set to 5e-3, 1e-6and 1e-5, respectively. Amongst these, the optimizer of the base network is monitored with its learning rate reduced when the loss plateaus. The loss checked for this plateau condition is the task loss on the validation set we constructed earlier and used as an unannotated data source. We train the step zero network for 120 epochs in the first stage and the 50 epochs in the next stage of the pipeline.
3 Experiments
3.1 Simulated dataset
The simulated data we use in our work is generated using a data generation process similar to (Che et al., 2019). The data generation process starts with density maps obtained directly from EMDB (Electron Microscopy Data Base) accessions or running situs package (Wriggers et al., 1999) on RCSB PDB (Protein Data Bank) files. In our experiments, density maps of four distinct classes (Ribosome, Proteosome, TRiC and Membrane) were obtained from EMDB accessions. These maps have a shape of with a voxel spacing of 1.368 nm. To simulate the missing wedge effect, the 3D structural data is projected on a series of 2D projection images with varying tilt angles based on the Maximum Wedge Angle hyperparameter specified. This projection data is then convolved with Contrast Transfer Function (CTF) and Modulation Transfer Function (MTF), adjusted by hyperparameters like Defocus and spherical aberration, for the simulation of the respective optical effects. At this stage, Gaussian Noise is added to the data to simulate the corresponding SNR level and the data is back-projected to obtain the 3D subtomogram. This 3D data is randomized by sampling hyperparameters uniformly from the following list, while in the specific case of the SNR, the sampling is done at a logarithmic scale.
SNR: 0.03 to 10
MWA: 0° to 50°
Dz: −12 to 0
Cs: 1.5 to 3.0
where MWA stands for Maximum Wedge Angle and Cs and Dz are hyperparameters for Spherical Abberation and Defocus, respectively. A sample from the simulated data is provided in Figure 4. Further, samples demonstrating the qualitative effect of these hyperparameters on the simulation is available in Supplementary Material.
Fig. 4.
2D slice visualization of simulated subtomograms
3.2 Experimental dataset
The real dataset we use in our experiments consists of 1051 subtomograms, which have been processed by template search and have been filtered manually from rat neuron tomograms (Guo et al., 2018). The data has a tilt angle range of −50 to +70 and is divided into four classes, namely ribosome, mitochondrial membrane, TRiC and single capped proteasome. A summary of the class-wise distribution of the data in terms of the number of subtomograms is available in Table 2. A figure containing slices of a subtomogram belonging to each class is available in Figure 3. This data is further split in a 1:9 ratio where the smaller part of the split is fed to the network as unlabeled real data for randomization and adaptation. Since this smaller subset of the data is passed as unlabeled and does not actively participate in reducing the classification loss, we reuse it as a validation set in our experiments.
Table 2.
Ablation studies in domain randomization on CB3D
| Modules used | Loss function | Macro Avg. |
|---|---|---|
| 0.72 | ||
| Wn | 0.72 | |
| Wd | 0.75 | |
| 0.77 |
The warp network was trained on reverse grad with the classifier. Bold entries represent the methodology described in the paper while the other entries in the corresponding tables either represent ablation experiments with different methodologies or represent comparisons with relevant algorithms.
Fig. 3.
2D slice visualization of experimental subtomograms
3.3 Ablation study
We perform ablation experiments for the warp network, as the feature discriminator part of our algorithm does not involve multiple modules. For our ablation studies, we use multiple neural networks varying in depth to demonstrate the efficacy of our methods on more than one network architecture. The network architectures we use along with their trainable parameter counts are listed in 1. Table 3 presents the performance of the warp module along with the performance of corresponding step zero networks (in terms of macro-average accuracy across four classes) while Figure 6 presents a visual demonstration of the effect of these modules. The presence of ‘(W)’ along with the neural network architecture used denotes that the model has been trained with the warp network, while the absence of this notation implies the performance recorded is that of the pre-trained step zero network. An increasing average accuracy while using the warp modules is expected here as we are not only using the domain randomization algorithm to fool the discriminator, but are also training the classifier to keep up with it. Table 4 presents more ablation studies using CB3D as the base network with modules zeroed out at places to demonstrate their individual contributions to the network. Table 4 also includes a study where we remove the domain discriminator and force the warp modules to train against the classification loss as proposed in Zakharov et al. (2019). Our ablation studies demonstrate the efficacy of each module in the warp network along with individual components of the loss function we use. Additional ablation experiments on simulated data only are provided in Supplementary Material.
Table 3.
Classwise accuracy comparison with step-zero networks
| Architecture | C 0 | C 1 | C 2 | C 3 | Avg. |
|---|---|---|---|---|---|
| CB3D | 0.70 | 0.63 | 0.61 | 0.91 | 0.71 |
| CB3D (W) | 0.71 | 0.75 | 0.69 | 0.92 | 0.77 |
| DSRF3D_v2 | 0.69 | 0.64 | 0.59 | 0.91 | 0.71 |
| DSRF3D_v2 (W) | 0.73 | 0.66 | 0.67 | 0.89 | 0.74 |
| DenseNet3D | 0.70 | 0.54 | 0.53 | 0.97 | 0.69 |
| DenseNet3D (W) | 0.84 | 0.73 | 0.45 | 0.96 | 0.74 |
| ResNet3D | 0.77 | 0.50 | 0.54 | 0.91 | 0.68 |
| ResNet3D (W) | 0.70 | 0.57 | 0.59 | 0.96 | 0.70 |
Bold entries represent the methodology described in the paper while the other entries in the corresponding tables either represent ablation experiments with different methodologies or represent comparisons with relevant algorithms.
Fig. 6.

Unsupervised domain adaptation results
Table 4.
Experimental data
| Class label | Subtomograms | Validation set | Test set |
|---|---|---|---|
| Ribosome (C0) | 80 | 8 | 72 |
| Proteasome (C1) | 386 | 38 | 348 |
| TRiC (C2) | 125 | 12 | 113 |
| Membrane (C3) | 460 | 46 | 414 |
| Total | 1051 | 104 | 947 |
4 Results
The extensive results of randomization and adaptation on the CB3D model are presented in Table 5. We compare our results with relevant domain adaptation methods like FSADA (Yu et al., 2021) and FADA (Motiian et al., 2017) in Table 6. While these are domain adaptation methods that make use of labels in limited amounts, we propose a completely unsupervised procedure of the same and outperform these methods by a large margin (7%). To further ensure a fair comparison with these methods, we train all of them on the same architecture as proposed by Yu et al. (2021). We also include t-SNE plots in Figure 5 that represent feature spaces extracted from M to visualize the reduction of domain shift between simulated and experimental data brought about by our training methodology. The t-SNE plots give a visual of how a domain shift is present between the features of the real and the simulated data. It further demonstrates how a reduction in the same has been achieved in the features extracted from our network.
Table 5.
Adaptation results on CB3D
| Label | Precision | Recall | F1_score | Support |
|---|---|---|---|---|
| Ribosome | 0.65 | 0.99 | 0.78 | 72 |
| Proteasome | 0.90 | 0.88 | 0.89 | 348 |
| TRiC | 0.83 | 0.78 | 0.80 | 113 |
| Membrane | 0.99 | 0.94 | 0.96 | 414 |
| Accuracy | 0.90 | 947 | ||
| Macro Avg. | 0.84 | 0.90 | 0.86 | 947 |
| Weighted Avg. | 0.91 | 0.90 | 0.90 | 947 |
Bold entries represent the methodology described in the paper while the other entries in the corresponding tables either represent ablation experiments with different methodologies or represent comparisons with relevant algorithms.
Table 6.
Accuracy comparison with other domain adaptation methods
| Methods | C 0 | C 1 | C 2 | C 3 | Avg. |
|---|---|---|---|---|---|
| FADA 3-shot (Motiian et al., 2017) | 0.62 | 0.70 | 0.66 | 0.91 | 0.72 |
| FADA 5-shot | 0.61 | 0.76 | 0.64 | 0.91 | 0.73 |
| FADA 7-shot | 0.78 | 0.64 | 0.62 | 0.91 | 0.73 |
| FSDA 3-shot (Yu et al., 2021) | 0.61 | 0.70 | 0.67 | 0.87 | 0.71 |
| FSDA 5-shot | 0.64 | 0.78 | 0.66 | 0.92 | 0.75 |
| FSDA 7-shot | 0.65 | 0.78 | 0.70 | 0.94 | 0.77 |
| Cryo-shift a | 0.75 | 0.85 | 0.83 | 0.95 | 0.84 |
Trained with our methodology on architecture from FSDA. Bold entries represent the methodology described in the paper while the other entries in the corresponding tables either represent ablation experiments with different methodologies or represent comparisons with relevant algorithms.
Fig. 5.
Individual affects of the warp modules on a subtomogram slice
5 Conclusion
The study of in situ macromolecular complexes enabled with cryo-electron tomography is essential for the analysis of various cellular processes. Shortcomings in the imaging procedure in conjunction with high structural complexity in tomograms make this analysis difficult to perform. The classification of subtomograms as sub-volumes of these tomograms forms an integral step in the structural recovery of macromolecules along with aiding in their analysis. Supervised classification methods, however, require large-scale annotated data that is extremely scarce. In this work, we propose an unsupervised domain adaptation and randomization framework which can help train neural networks completely on simulated data and unlabeled experimental data. We first train a base network on simulated data and then use a network-driven domain randomization framework to help the network make better predictions on the experimental dataset. Additionally, we make use of a multi-adversarial domain adaptation module where we help the classifier generate similar features for both real and simulated data, thus mitigating the domain shift between the real and the simulated data features and helping the model classify better in both of these domains. Our work can be further used for automated annotation of subtomograms after being trained on similar simulated data, thus forming an important step in the completely autonomous recognition and structural recovery of subtomograms with the help of neural networks for better analysis of macromolecules and their interactions.
We find our work can be improved upon with future works aimed at further development of simulation algorithms to bridge the domain gap between real and simulated data. Unsupervised deep learning-based algorithms for the analysis of genomic data (Liu et al., 2021b) have made considerable progress with the help of generative networks. A possibility lies in the use of novel conditional generative networks (Liu et al., 2021a; Vahdat et al., 2021) to generate class-specific cryo-et data for improving upon the algorithms demonstrated in this work.
Supplementary Material
Acknowledgements
The authors thank Dr Qiang Guo for sharing with us experimental rat-neuron tomograms that the authors make use of as real data.
Funding
This work was supported in part by U.S. National Institutes of Health (NIH) [R01GM134020 and P41GM103712], U.S. National Science Foundation (NSF) [DBI-1949629 and IIS-2007595], AMD COVID-19 HPC Fund and the Mark Foundation For Cancer Research [19-044-ASP]. X.Z. was supported in part by a fellowship from Center of Machine Learning and Health at Carnegie Mellon University.
Conflict of Interest: none declared.
Contributor Information
Hmrishav Bandyopadhyay, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata 700032, India.
Zihao Deng, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Leiting Ding, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Sinuo Liu, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Mostofa Rafid Uddin, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Xiangrui Zeng, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Sima Behpour, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
Min Xu, Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA.
References
- Bartesaghi A. et al. (2008) Classification and 3D averaging with missing wedge correction in biological electron tomography. J. Struct. Biol., 162, 436–450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beck M. et al. (2009) Visual proteomics of the human pathogen leptospira interrogans. Nat. Methods, 6, 817–823. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Best C. et al. (2007) Localization of protein complexes by pattern recognition. Methods Cell Biol., 79, 615–638. [DOI] [PubMed] [Google Scholar]
- Briggs J.A. (2013) Structural biology in situ—the potential of subtomogram averaging. Curr. Opin. Struct. Biol., 23, 261–267. [DOI] [PubMed] [Google Scholar]
- Che C. et al. (2018) Improved deep learning-based macromolecules structure classification from electron cryo-tomograms. Mach. Vis. Appl., 29, 1227–1236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Che C. et al. (2019) Domain randomization for macromolecule structure classification and segmentation in electron cyro-tomograms. In: 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) San Diego, CA, USA. IEEE. pp. 6–11.
- Chen M. et al. (2017) Convolutional neural networks for automated annotation of cellular cryo-electron tomograms. Nat. Methods, 14, 983–985. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Y. et al. (2014) Autofocused 3D classification of cryoelectron subtomograms. Structure, 22, 1528–1537. [DOI] [PubMed] [Google Scholar]
- Du X. et al. (2021) Active learning to classify macromolecular structures in situ for less supervision in cryo-electron tomography. Bioinformatics, 37, 2340–2346. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ganin, Y., and Lempitsky, V. (2015, June). Unsupervised domain adaptation by backpropagation. In International conference on machine learning (pp. 1180-1189). PMLR.
- Gao S. et al. (2020) Dilated-densenet for macromolecule classification in cryo-electron tomography. In: International Symposium on Bioinformatics Research and Applications, Moscow, Russia. Springer, pp. 82–94. [DOI] [PMC free article] [PubMed]
- Guo Q. et al. (2018) In situ structure of neuronal C9ORF72 poly-GA aggregates reveals proteasome recruitment. Cell, 172, 696–705. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hara K. et al. (2018) Can spatiotemporal 3D CNNs retrace the history of 2D CNNs and imagenet? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, Utah. pp. 6546–6555.
- Kunz M. et al. (2015) M-free: mask-independent scoring of the reference bias. J. Struct. Biol., 192, 307–311. [DOI] [PubMed] [Google Scholar]
- Lin R. et al. (2019) Adversarial domain adaptation for cross data source macromolecule in situ structural classification in cellular electron cryo-tomograms. Bioinformatics, 35, i260–i268. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu Q. et al. (2021a) Density estimation using deep generative neural networks. Proc. Natl. Acad. Sci. USA, 118, e2101344118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu Q. et al. (2021b) Simultaneous deep generative modelling and clustering of single-cell genomic data. Nat. Mach. Intell., 3, 536–544. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liu S. et al. (2019) Semi-supervised macromolecule structural classification in cellular electron cryo-tomograms using 3D autoencoding classifier. In: BMVC, Cardiff University, UK.
- Liu S. et al. (2020) Efficient cryo-electron tomogram simulation of macromolecular crowding with application to SARS-COV-2. In: 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE. pp. 80–87.
- Lučić V. et al. (2013) Cryo-electron tomography: the challenge of doing structural biology in situ. J. Cell Biol., 202, 407–419. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Motiian S. et al. (2017) Few-shot adversarial domain adaptation. Adv. Neural Inf. Process. Syst., 30, 6670–6680. [Google Scholar]
- Pei L. et al. (2016) Simulating cryo electron tomograms of crowded cell cytoplasm for assessment of automated particle picking. BMC Bioinformatics, 17, 1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pei Z. et al. (2018) Multi-adversarial domain adaptation. Proc. AAAI Conf. Artif. Intell., 32. [Google Scholar]
- Vahdat A. et al. (2021) Score-based generative modeling in latent space. arXiv preprint arXiv, 2106, 05931.
- Wriggers W. et al. (1999) Situs: a package for docking crystal structures into low-resolution maps from electron microscopy. J. Struct. Biol., 125, 185–195. [DOI] [PubMed] [Google Scholar]
- Xu M. et al. (2012) High-throughput subtomogram alignment and classification by Fourier space constrained fast volumetric matching. J. Struct. Biol., 178, 152–164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xu M. et al. (2017) Deep learning-based subdivision approach for large scale macromolecules structure recovery from electron cryo tomograms. Bioinformatics, 33, i13–i22. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yu L. et al. (2021) Few shot domain adaptation for in situ macromolecule structural classification in cryoelectron tomograms. Bioinformatics, 37, 185–191. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zakharov S. et al. (2019) Deceptionnet: network-driven domain randomization. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, South Korea. pp. 532–541.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.




