Skip to main content
iScience logoLink to iScience
. 2020 Sep 29;23(10):101601. doi: 10.1016/j.isci.2020.101601

MitoSegNet: Easy-to-use Deep Learning Segmentation for Analyzing Mitochondrial Morphology

Christian A Fischer 1,2,3, Laura Besora-Casals 1, Stéphane G Rolland 1, Simon Haeussler 1, Kritarth Singh 4, Michael Duchen 4, Barbara Conradt 1,2,4,, Carsten Marr 3,5,∗∗
PMCID: PMC7554024  PMID: 33083756

Summary

While the analysis of mitochondrial morphology has emerged as a key tool in the study of mitochondrial function, efficient quantification of mitochondrial microscopy images presents a challenging task and bottleneck for statistically robust conclusions. Here, we present Mitochondrial Segmentation Network (MitoSegNet), a pretrained deep learning segmentation model that enables researchers to easily exploit the power of deep learning for the quantification of mitochondrial morphology. We tested the performance of MitoSegNet against three feature-based segmentation algorithms and the machine-learning segmentation tool Ilastik. MitoSegNet outperformed all other methods in both pixelwise and morphological segmentation accuracy. We successfully applied MitoSegNet to unseen fluorescence microscopy images of mitoGFP expressing mitochondria in wild-type and catp-6ATP13A2 mutant C. elegans adults. Additionally, MitoSegNet was capable of accurately segmenting mitochondria in HeLa cells treated with fragmentation inducing reagents. We provide MitoSegNet in a toolbox for Windows and Linux operating systems that combines segmentation with morphological analysis.

Subject Areas: Cell Biology, Bioinformatics, Automation in Bioinformatics, Artificial Intelligence

Graphical Abstract

graphic file with name fx1.jpg

Highlights

  • MitoSegNet is a deep learning model for quantifying mitochondrial morphology

  • MitoSegNet outperforms other mitochondria segmentation methods

  • MitoSegNet was used to identify previously unreported morphological differences

  • MitoSegNet successfully segments mitochondria in HeLa cells


Cell Biology; Bioinformatics; Automation in Bioinformatics; Artificial Intelligence

Introduction

Cellular organelles are integral to eukaryotic cells and their functions. One organelle that has always been of particular interest is the mitochondrion, which plays an essential role in several metabolic pathways including that of Adenosine triphosphate (ATP). Mitochondria are often represented as static, bean-shaped organelles but actually form highly dynamic ‘tubular’ networks that often undergo changes in distribution and morphology (Tilokani et al., 2018; Chan, 2020). The steady-state morphology of mitochondria in a cell is a result of a balance between two opposing processes, mitochondrial fusion and fission. Changes in this balance result in changes in mitochondrial morphology. It has been shown that changes in morphology allow mitochondria to respond to metabolic or environmental stresses, while maintaining homeostasis (Tondera et al., 2009; Rolland et al., 2013; Wai and Langer, 2016). While partially damaged mitochondria can be rescued by exchanging their contents with functional mitochondria through mitochondrial fusion, mitochondrial fission enables the removal of damaged mitochondria and can also facilitate apoptosis during increased levels of cellular stress (Pernas and Scorrano, 2016). Mitochondrial fusion and fission are regulated by a conserved family of dynamin-related GTPases and have been well studied in Caenorhabditis elegans (van der Bliek et al., 2017). In C. elegans, the membrane anchored dynamin-related GTPases FZO-1MFN and EAT-3OPA1 are required for the fusion of the outer- and inner mitochondrial membranes, respectively. The loss of function of either of these two proteins results in mitochondria with a ‘fragmented’ morphology (Breckenridge et al., 2008; Ichishita et al., 2008; Kanazawa et al., 2008; Tan et al., 2008; Rolland et al., 2009).

Mitochondrial fission in C. elegans is mediated by DRP-1DRP−1, a cytosolic dynamin-related GTPase. Depletion of DRP-1 has been shown to result in mitochondria with an ‘elongated’ morphology (Labrousse et al., 1999). Mutations in the human orthologs of the genes encoding these proteins have been associated with several diseases, including neurodegenerative diseases (Chan, 2020). For this reason, understanding mitochondrial fusion and fission is not only an important basic biological question but is critical for our ability to understand the pathology of these diseases and to develop novel therapeutics to treat them. However, such studies have been hindered by the fact that is difficult to assess mitochondrial morphology in different genetic backgrounds or physiological conditions in an unbiased and quantitative manner. Specifically, the diversity of shapes among mitochondria (elongated, fragmented, tubular, as well as ‘mixed’ morphologies) poses a challenge to the automated quantification of mitochondrial morphology. For this reason, researchers often resorted to the use of a simple qualitative assessment of mitochondrial morphology. As a result, subtle differences in morphology and, hence, phenotypes are often not detected. To analyze mitochondrial morphology, for example, in C. elegans, mitochondria are labeled using either a mitochondria-specific fluorescent dye (such as TMRE) or a transgene expressing a mitochondrial-targeted GFP (mitoGFP) (Regmi and Rolland, 2017) and a fluorescent microscopy image is acquired. In order to quantify the mitochondrial morphology in an automated and unbiased manner, the next critical step is to segment the objects in the fluorescent image. The simplest form of image segmentation is thresholding, which is only successful if features are well separated and their intensities vary considerably from the background (Torborg and Feller, 2004). These requirements are seldom met in live cell imaging due to autofluorescence, noise or fluctuating intensities. Thresholding segmentation can be improved through the prior application of feature enhancement algorithms based on intensity derived features, such as the Difference-of-Gaussians (DoG), Determinant of Hessian (Sato et al., 1998) or Laplacian-of-Gaussian (LoG), which are also known as blob detectors. DoG is used to enhance the visibility of edges by removing high frequency information but at the cost of reducing the overall image contrast, while LoG is useful for detecting edges that appear at different image scales or degrees of focus (Marr and Hildreth, 1980; Lindeberg, 1994). Curvilinear structures (such as nerve fibers or blood vessels) can be segmented using the eigenvalues of a Hessian matrix, with which one can calculate the object curvature. There are a many other methods available used in segmentation workflows, such as morphological filtering (dilation, erosion, etc.), region accumulation (watershed transform), deformable model fitting (active contour model) and machine learning (k-means clustering, random forest, etc.) (Meijering, 2012). Most of these methods can now be implemented by biologists through free and opensource tools such as Fiji (Schindelin et al., 2012), CellProfiler (McQuin et al., 2018) or Ilastik (Berg et al., 2019).

All of these segmentation methods have shown varying degrees of success depending on the images they were supposed to segment (de Boer et al., 2015; Li et al., 2015; Akram et al., 2017; Berg et al., 2019). However, with rising image complexity as well as a decreased signal-to-noise ratio, most of the methods perform poorly. For such cases, the only option in the past was manual segmentation, which is highly laborious and introduces a varying degree of bias on each labeled image. With the recent emergence of deep learning and in particular the development of convolutional neural networks (CNNs) (LeCun et al., 1989; Krizhevsky et al., 2012) automated approaches that perform these tasks with human accuracy have become available. CNNs were inspired by the research of Hubel and Wiesel on the primary visual cortex of cats (Wiesel and Hubel, 1963). CNN's can classify data based on convolution and pooling operations. Convolution describes the extraction of features from an image by sliding filters across the image and generating feature maps. Pooling reduces the dimensionality of each feature map, while retaining the most important information. It also reduces the number of network parameters, prevents overfitting, and makes the network invariant to small distortions in the input image (Scherer et al., 2010). Through the successive and repetitive application of convolution and pooling, CNNs are capable of classifying highly complex images with great accuracy (Szegedy et al., 2014). To perform semantic segmentation, which is the assignment of a class label to each pixel, one must use a fully convolutional neural network (FCNN) (Long et al., 2014). A popular FCNN in the biological community is the U-Net that was specifically developed for biomedical image segmentation (Ronneberger et al., 2015). It has been successfully applied to many different biomedical image segmentation tasks and yields good results with only a few hand-segmented images (Chlebus et al., 2018; De Fauw et al., 2018; Stember et al., 2018).

In this study, we trained a U-Net, which we named Mitochondrial Segmentation Network (MitoSegNet), to learn how to segment mitochondria in adult C. elegans body wall muscle cells, compared its performance and tested its generalizability in biologically relevant applications that demonstrate C. elegans animals carrying a loss-of-function mutation in the gene catp-6ATP13A2 exhibit a previously unreported mitochondrial morphology phenotype. We also show that MitoSegNet can be successfully used to analyze mitochondrial morphology in HeLa cells.

Results

The MitoSegNet Model

The MitoSegNet model was generated by training a modified U-Net with a training set of 12 1300 × 1030 pixel fluorescent microscopy, maximum-intensity projection images, depicting mitochondria in body wall muscle cells of adult C. elegans worms (mitochondria were visualized using a transgene expressing mitochondrial matrix-targeted GFP under the control of a body wall muscle-specific promoter (Pmyo3::mitoGFP)) (Figure 1A and Methods). Our U-Net modification entails the removal of dropout layers at the end of the contracting pathway and instead placing batch normalization layers after every convolutional layer prior to ReLU activation in the contracting pathway. This modification decreased the amount of necessary training time. Each image was split into 4 overlapping tiles. For each tile, 80 augmented copies were generated for training the model. A cross validation was performed to estimate the performance of the MitoSegNet on an unseen test set and to compare it against other segmentation methods (Figure 1B).

Figure 1.

Figure 1

Training the MitoSegNet Model and Using It with the MitoS Tool

(A) The original training data is comprised 12 raw images and the appendant hand-generated ground truth images. Each image is split into 4 overlapping tiles of equal length. For each tile, a weight map is generated and subsequently all three set of tiles (raw images, ground truth, and weight maps) are augmented 80 times, increasing the size of the training data to 3,840 image tiles. Prior to training, the augmented training data is split into training (80%) and validation data (20%). The pretrained MitoSegNet model can now be used to segment new images of mitochondria.

(B) We performed a cross validation for which 12 separate MitoSegNet models were trained each with 11 images, excluding one image that was later used to test the prediction accuracy against other segmentation methods.

Visual Comparison of Segmentation Performance

To qualitatively evaluate the performance of the MitoSegNet, we compared the predicted segmentations against manually segmented ground truth in an unseen test set. The same procedure was repeated for four other segmentation methods. We considered three classical feature enhancement methods (Gaussian, Hessian, and Laplacian) followed by different thresholding algorithms, all implemented in ImageJ/Fiji. The fourth method is the machine-learning segmentation tool Ilastik (Kreshuk and Zhang, 2019). The Gaussian, Hessian, Laplacian, and Ilastik methods failed to consistently prevent false positive and/or false negative segmentation on all phenotypes (Figure 2). The Gaussian segmentation produced large sections of false positive predictions in the mixed and tubular phenotype (indicated by yellow arrows). The Hessian and Laplacian segmentation largely avoided false positive predictions but instead often failed to recognize mitochondria, resulting in false negative segmentations in the elongated, mixed, and tubular phenotype (and fragmented for the Laplacian segmentation) (Figure 2). The Ilastik-based segmentation produced only very little false negative predictions but like the Gaussian segmentation, predicted large amounts of false positives in all but the fragmented phenotype. The MitoSegNet segmentation drastically reduces the amount of false negative or false positive segmentation when compared to the other methods and yielded consistent results across all different phenotypes (Figure 2).

Figure 2.

Figure 2

MitoSegNet Visually Outperforms Feature-Based Segmentation Approaches

On the left side, four images of elongated, fragmented, tubular, and mixed C. elegans mitochondria and their respective ground truth are shown. The masks on the right show the results of MitoSegNet and the four segmentation methods applied to each image, displaying the false negative segmentation in red, the false positive segmentation in turquoise, the true negative segmentation as black, and the true positive segmentation as white. The yellow arrows indicate areas in which false segmentation occured. The scale bar is 5 μm.

Quantitative Comparison of Segmentation Performance

To compare the methods' performance more quantitatively, we evaluated the pixelwise segmentation accuracy using the dice coefficient (Taha and Hanbury, 2015). The MitoSegNet outperforms the feature-based and Ilastik-generated segmentations (Figure 3A) with a median dice coefficient of 0.89 and a lower and upper 95% confidence interval of 0.87 and 0.91 (N = 12) significantly (p = 5.11∗10−5, Kruskal-Wallis test). However, pixelwise accuracy as measured by the dice coefficient does not necessarily guarantee correct prediction of morphology (Figure 3B). Because segmented images in biology are often used for morphological quantification (de Boer et al., 2015; Abdolhoseini et al., 2019; Orozco-Fuentes et al., 2019), we assessed the morphological accuracy with two other approaches. The single object shape deviation per object was measured for five shape descriptors (area, eccentricity, aspect ratio, perimeter, and solidity) and averaged over 12 images (Figure 3C). The MitoSegNet with a median average fold deviation of 1.09 and a lower and upper 95% confidence interval of 1.07 and 1.12 (N = 60) outperforms all other methods in the accurate prediction of single object morphology (p = 7.4∗10−10, Kruskal-Wallis test) (Figure 3D). Because the single object shape deviation method does not consider false negative predictions, we also compared all segmented objects in ground truth and prediction. For each image and each of the five object descriptors, the energy distance between the ground truth and predicted distributions was calculated (Figure 3E). Due to the different value ranges among the descriptors the values were normalized prior statistical analysis. The MitoSegNet segmentation achieves a median normalized energy distance of 0.20 with a lower and upper 95% confidence interval of 0.16 and 0.23 (N = 60) and again statistically outperforms all other non-deep learning segmentation methods (p = 3.3∗10−18, Kruskal-Wallis test) (Figure 3F).

Figure 3.

Figure 3

MitoSegNet's Pixelwise Accuracy Outperforms Non-Deep Learning Segmentation Methods

(A) The average dice coefficient achieved with the MitoSegNet is significantly higher than with the four other segmentation approaches. The data was statistically evaluated by using the Kruskal-Wallis test followed by a Dunn's multiple comparisons test.

(B) The dice coefficient is limited as a predictor of morphological segmentation performance: A dice coefficient close to 1.0 does not guarantee correct prediction of morphology. Contrary, a low dice coefficient does not rule out an accurate shape prediction. Ground truth segmentation is shown in orange, and the predicted segmentation in green.

(C) To gain insight into how accurately the shape of ground truth objects is predicted, five object shape descriptors (area, eccentricity, aspect ratio, perimeter, and solidity) are compared by calculating the fold deviation. Predicted objects that correspond to more than one ground truth object (or vice versa) are excluded from this analysis.

(D) The MitoSegNet shows the lowest average fold deviation between predicted and ground truth object shape descriptors. The data were statistically evaluated by first testing for normality using D'Agostino's K-squared test and then subsequently using the Kruskal-Wallis test followed by a Dunn's multiple comparisons test. N = 60.

(E) To determine the total morphological prediction accuracy, the same five shape descriptors were measured for each image. The descriptor distributions in the ground truth and predicted images were statistically evaluated for differences by calculating the energy distances between predicted and ground truth distribution. The energy distances for each shape descriptor and image were normalized prior to statistical analysis.

(F) The MitoSegNet shows the lowest normalized energy distance, statistically outperforming all other segmentation approaches. The data were first tested for normality using the D'Agostino's K-squared. After determining that all distributions were non-parametric, a Kruskal-Wallis test was used followed by a Dunn's multiple comparisons test. N = 60. ∗p < 0.05, ∗∗0.001 < p < 0.01, ∗∗∗p < 0.001 for (A), (D) and (F).

Comparison of Mitochondrial Morphology between Catp-6ATP13A2 Mutant and Wild-type

To evaluate the applicability of the MitoSegNet on a different, unseen set of images, we decided to use the MitoSegNet to determine whether a loss-of-function mutation of the gene catp-6ATP13A2, ok3473 (hereafter referred to as catp-6(lf)), causes a mitochondrial morphology phenotype. catp-6ATP13A2 encodes a member of the family of P-type ATPases, which transport various compounds across membranes using ATP hydrolysis as energy source (Moller et al., 1996; Lambie et al., 2013; Anand et al., 2020). In addition, catp-6ATP13A2 is the human ortholog of ATP13A2, mutations of which leads to Kufor-Rakeb syndrome, a form of inherited juvenile-onset Parkinsons disease (Ramirez et al., 2006; Di Fonzo, Chien et al., 2007). No abnormal differences in mitochondrial morphology has so far been reported for the catp-6(lf). Consistent with this, upon brief visual inspection, no obvious differences in mitochondrial morphology are noticeable (Figure 4A). We applied the MitoSegNet (Figures 4A) to 19 fluorescence microscopy images of each genotype and subsequently analyzed the data with the MitoA tool (Figure 4B). Segmentation masks (Figure 4A) visually matched the raw images closely and subsequent quantification revealed a statistically significant morphological difference between mitochondria in wild-type and catp-6(lf) mutants. Compared to wild-type, mitochondria are significantly thinner and longer in catp-6(lf) mutants, as determined by the average minor (p = 0.047, independent two-sample t test) and major axis length (p = 0.029, independent two-sample t test) (Figure 4B). Furthermore, the average mitochondrial area is larger (p = 0.00039, independent two-sample t test) and the perimeter is longer (p = 0.043, independent two-sample t test) in wild-type compared to catp-6(lf) mutants (Figure 5B). Excessive mitochondrial fission (i.e. mitochondrial fragmentation) as a cause for these observations can be excluded since the numbers of mitochondria in wild-type and catp-6(lf) mutants are similar (p = 0.56, independent two-sample t test) (Figure 4B). Differences were also found in terms of mitochondrial branch morphology (Figure 4C). While the average mitochondrial branch length in catp-6(lf) mutants is larger than in wild-type (p = 0.01, independent two-sample t test), the average number of mitochondrial branches was found to be significantly smaller in catp-6(lf) mutants (p = 0.009, Mann-Whitney U test). Altogether these results indicate that catp-6(ok3473) causes a previously unreported mitochondrial morphology phenotype that cannot be detected by simple visual inspection.

Figure 4.

Figure 4

Wild-type and Catp-6 Mutant Mitochondria Show Significant Morphological Differences

(A) Visual comparison of catp-6 mutant and wild-type mitochondrial morphology. Raw images are at the top and MitoSegNet model segmentations at the bottom. The scale bar is 3 μm.

(B) Mitochondrial shape descriptor comparison.

(C) Mitochondrial branch descriptor comparison. Average area, minor and major axis length (see scheme for explanation), perimeter, number of mitochondria, average branch length, and number of branches were measured in segmented images of wild-type and catp-6 mutant mitochondria. ∗p < 0.05, ∗∗0.001 < p < 0.01, ∗∗∗p < 0.001 using the Mann-Whitney U test. N = 19.

Figure 5.

Figure 5

MitoSegNet Model Segmentation and Morphological Quantification Can Be Applied to Mitochondria of Mammalian Cells

Comparing untreated HeLa cells and HeLa cells treated with oligomycin or antimycin for a duration of 3 h.

(A) Visual comparison of untreated and treated mitochondrial morphology. Raw images are at the top and MitoSegNet model segmentations at the bottom. The scale bar is 2.5 μm.

(B) Average area, eccentricity, perimeter, and branch length of mitochondria were measured in segmented images of treated and untreated mitochondria. ∗p < 0.05, ∗∗∗p < 0.001 using an independent two-sample t test. N = 8.

MitoSegNet Segmentation of Mitochondria in HeLa Cells

To test the generalizability of our model even further, we applied the MitoSegNet to segment 8 confocal microscopy images depicting mitochondria in HeLa cells (Figure 5A). The fragmentation of mitochondria in HeLa cells treated with oligomycin and antimycin for 3 hr was captured in the segmentation both visually and quantitatively. As expected, the average mitochondrial area is significantly larger in untreated cells compared to treated HeLa cells (p = 0.0068, independent two-sample t test) (Figure 5B). The average eccentricity is lower for the fragmented mitochondria compared to the untreated mitochondria, indicating a more circular shape (p = 1.32∗10−8, independent two-sample t test) (Figure 5B). The average perimeter distribution reflects a similar pattern as found for the area, showing the fragmented mitochondria to have a smaller perimeter on average (p = 0.00037, independent two-sample t test) (Figure 5B). The average branch length is also significantly smaller in the treated mitochondria compared to the untreated mitochondria (p = 1.30∗10−5, independent two-sample t test) (Figure 5B).

The MitoS Segmentation and MitoA Analysis Tool

To enable non-experts, we implemented the MitoSegNet in an easy-to-use tool, the MitoS segmentation tool, a Python-based, standalone executable. MitoS can be executed in a basic mode, which utilizes the pretrained MitoSegNet for segmentation of mitochondria and allows us to easily apply the model without prior deep learning experience (Figure 1). We applied the MitoS image segmentation using the pretrained MitoSegNet on two different systems. For all cases 10 images of each 1300 × 1030 px size (8-bit) were segmented. The MitoS GPU segmentation was run using an NVIDIA GeForce GTX 960M and NVIDIA TITAN X and segmentation took 65 and 15 s, respectively. Segmentation using the MitoS CPU version was performed on an Intel(R) Core(TM) i7-6700HQ CPU and a system using four Intel(R) Xeon(R) CPU E5-2680 v4 processors and lasted 7.5 min and 65 s, respectively. The basic mode also includes a fine-tuning module, which allows us to optimize the pretrained on new images (Figure S1A). The MitoS advanced mode can be used if other structures besides mitochondria should be segmented or if the user wishes to build a self-configured deep learning segmentation model (Figure S1B). The MitoA analyzer is a separate Python-based, standalone executable tool that can be run after successful segmentation for quantification and visualization of potential morphological differences (Figure S3). It measures ten different morphological and three intensity-based features for each object and summary statistics for all object features per image are generated. The tables of two or multiple samples containing these summary statistics can then be subjected to hypothesis testing, visualization, and correlation analysis. The MitoS and MitoA tools require no installation and no prerequisite installations (such as frameworks), and they are available for both Windows and Linux.

Discussion

We present MitoSegNet, a segmentation model that exploits the power of deep learning to address the challenging problem of accurate mitochondria segmentation. We show that the MitoSegNet outperforms feature-based, non-deep learning-based algorithms and that it is generalizable to unseen images from C. elegans and mammalian cells.

MitoSegNet Model Segmentation Performance

While the superior visual and quantitative performance of the MitoSegNet model segmentation might not come as a surprise to researchers acquainted with the capabilities of deep learning-based segmentation, we believe these results to be interesting to researchers who commonly use feature-based segmentation methods. For accuracy evaluation, we did not rely only on pixel-based accuracy, as we found this to be an insufficient measure of morphological accuracy but extended our analysis. Our single object shape comparison as well as the calculation of energy distances for five different feature descriptors per image demonstrate that the MitoSegNet model segmentation yields the best morphological accuracy compared with commonly applied segmentation methods.

Comparison of Mitochondrial Morphology between Catp-6ATP13A2 Mutants and Wild-type

The visual comparison of mitochondrial morphology in wild type and catp-6(lf) mutants did not reveal any obvious differences in morphology. In both cases, mitochondria appeared to be largely tubular. However, the quantitative analysis using the MitoSegNet revealed that average area, perimeter, and minor axis length of mitochondria in catp-6(lf) mutants are smaller than in wild-type. Since the average number of mitochondria per cell is the same as in wild-type, we concluded that catp-6(lf) causes a reduction in overall mitochondrial mass. This is consistent with observations of Anand et al., who found that oxygen consumption is lower in C. elegans catp-6(lf) mutants than in wild type (Anand et al., 2020). In mammalian cells, mutation of ATP13A2 has been found to cause impairment of mitochondrial function and induce mitochondrial fragmentation (Gusdon et al., 2012; Ramonet et al., 2012). However, this is associated with increased mitochondrial mass due to inefficient autophagic clearance (Grünewald et al., 2012). The differences between the mutant phenotypes observed in C. elegans and mammalian cells is likely to reflect the multiple roles of ATP13A2 and CATP-6 transport substrates (polyamines) in maintaining mitochondrial function (van Veen et al., 2020). Further research on properties of catp-6 deficient mitochondria, such as membrane potential or levels of reactive oxygen species, might uncover the cause for the observed differences in mitochondrial morphology.

Application of the MitoSegNet to Mitochondria in HeLa Cells

Although the pretrained model was generated with standard fluorescence microscopy images, depicting mitoGFP-labeled mitochondria in C. elegans, the same pretrained model was able to generate visually accurate segmentations of mitoRFP labeled mitochondria in HeLa cells using a laser scanning confocal microscope. This demonstrates the high robustness and generalization capabilities of our pretrained MitoSegNet model and that it can be used for segmentation of mitochondria in organisms other than C. elegans. Furthermore, our MitoSegNet Analysis tool quantitatively confirmed the morphological differences of mitochondria between untreated HeLa cells and HeLa cells treated with oligomycin or antimycin.

MitoSegNet Model Architecture

Although the MitoSegNet architecture is largely based on the U-Net, through testing various changes in the original architecture, we found that the validation dice coefficient as well as the validation loss improved upon removing the dropout layers and instead placing a batch normalization layer (Ioffe and Szegedy, 2015) after every convolution layer in the contracting pathway. Interestingly, a recent study found that the combined usage of batch normalization followed by dropout (forming an independent component layer) stabilized the training process, increased convergence speed, and improved the convergence limit (Chen et al., 2019). It would require further testing to find out if the usage of an independent component layer would improve the current MitoSegNet performance even further.

MitoS and MitoA Tools

Most deep learning applications in the field of biological image segmentation were created for the purpose of 2D cell segmentation (Chen et al., 2017; Al-Kofahi et al., 2018; Falk et al., 2019; Kusumoto and Yuasa, 2019), while organelle-specific deep learning applications are scarce. Although most tools allow the user to retrain available 2D cell segmentation models to segment other biological structures of interest, this often requires computer science-related skills, such as familiarity with programming languages, shell interaction or knowledge on how to install various deep learning frameworks. One of the main motivations behind MitoSegNet and the MitoS and MitoA toolbox was to make deep learning segmentation accessible to researchers that do not have an extensive background in computer science or deep learning. The MitoS tool can be run without installation. The simple graphical user interface allows users to quickly navigate the MitoS and MitoA tools. The MitoS basic mode also comes with a fine-tuning module that allows researchers that would like to segment other organelles or images taken under different conditions than those used for training the MitoSegNet model. Since the subsequent step after segmentation is usually the analysis, we included the MitoA tool to save researchers the time to look up appropriate analysis tools and instead be able to quickly obtain potentially interesting insights.

Conclusion

The MitoSegNet has been shown to outperform both conventional feature-based and machine-learning-based segmentation of mitochondria. The pretrained model can be easily applied to new 2D microscopy images of mitochondria through the usage of the MitoS tool, which is available for both standard and high-end Windows and Linux systems. Successfully segmented images of mitochondria can be subjected to quantification, statistical testing, and visualization with the MitoA tool.

Limitations of the Study

The MitoSegNet model used for segmentation of images depicting fluorescently labeled mitochondria in C. elegans and HeLa cells was trained with 12 pairs of raw images and the appendant hand-labeled ground truth images. Although both the visual and quantitative segmentation accuracy was shown to be high in this study, there remains a bias which is based on the two annotators who generated 6 ground truth images each. To reduce the ground truth bias and increase the generalizability of the MitoSegNet model, more images labeled by different annotators can be added. Furthermore, image sections in which single mitochondria were not clearly distinguishable due to optical constraints of the microscope used to generate the image, made it difficult for the annotator to create labeled images that accurately represent the ground truth. This uncertainty introduced to the MitoSegNet model can be decreased by adding images recorded with higher resolution, thus reducing the sections in which such visually indistinguishable mitochondria exist.

Resource Availability

Lead Contact

Further information and requests for resources should be directed to and will be fulfilled by the Lead Contact, Carsten Marr (carsten.marr@helmholtz-muenchen.de).

Materials Availability

Images used for training and testing the model are available upon request.

Data and Code Availability

The software documentation for the MitoS and MitoA tool can be found at https://github.com/mitosegnet. The MitoSegNet segmentation model, the MitoA analysis and MitoS segmentation tool (GPU/CPU) for Linux and Windows are available at https://zenodo.org/search?page=1&size=20&q=mitosegnet. The Python code used for generating the figures displayed in the manuscript is available at https://github.com/MitoSegNet/MitoSegNet_AccuracyTesting_Manuscript.

Methods

All methods can be found in the accompanying Transparent Methods supplemental file.

Acknowledgments

We thank E. Lambie, H. Hartz and members of the Conradt and Marr lab for comments on the project development, the manuscript, and testing the MitoS and MitoA tools. We also thank Iva Dzhilyanova for testing the MitoS tool on multiple systems. This work was supported by the Deutsche Forschungsgemeinschaft (Center for Integrated Protein Science Munich [CIPSM; EXC 114], CO204/6-1 and CO204/9-1). C.M. has received funding from the European Research Council (ERC) under the European Union's Horizon 2020 research and innovation program (Grant agreement No. 866411). Some strains were provided by the CGC, which is funded by NIH Office of Research Infrastructure Programs (P40 OD010440).

Author Contributions

C.A.F. designed and conducted the computational experiments, performed the data analysis, and wrote the paper. B.C. designed the biological experiments and wrote the paper. C.M. designed the computational experiments and wrote the paper. S.G.R., S.H., L.B.C., M.D. and K.S. designed and conducted the biological experiments.

Declaration of Interests

None.

Published: October 23, 2020

Footnotes

Supplemental Information can be found online at https://doi.org/10.1016/j.isci.2020.101601.

Contributor Information

Barbara Conradt, Email: b.conradt@ucl.ac.uk.

Carsten Marr, Email: carsten.marr@helmholtz-muenchen.de.

Supplemental Information

Document S1. Transparent Methods and Figures S1–S5
mmc1.pdf (1.5MB, pdf)

References

  1. Abdolhoseini M., Kluge M.G., Walker F.R., Johnson S.J. Segmentation, tracing, and quantification of microglial cells from 3D image stacks. Sci. Rep. 2019;9:8557. doi: 10.1038/s41598-019-44917-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Akram F., Garcia M.A., Puig D. Active contours driven by difference of Gaussians. Sci. Rep. 2017;7:14984. doi: 10.1038/s41598-017-14502-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Al-Kofahi Y., Zaltsman A., Graves R., Marshall W., Rusu M. A deep learning-based algorithm for 2-D cell segmentation in microscopy images. BMC Bioinformatics. 2018;19:365. doi: 10.1186/s12859-018-2375-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Anand N., Holcom A., Broussalian M., Schmidt M., Chinta S.J., Lithgow G.J., Andersen J.K., Chamoli M. Dysregulated iron metabolism in C. elegans catp-6/ATP13A2 mutant impairs mitochondrial function. Neurobiol. Dis. 2020;139:104786. doi: 10.1016/j.nbd.2020.104786. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Berg S., Kutra D., Kroeger T., Straehle C.N., Kausler B.X., Haubold C., Schiegg M., Ales J., Beier T., Rudy M. Ilastik: interactive machine learning for (bio)image analysis. Nat. Methods. 2019;16:1226–1232. doi: 10.1038/s41592-019-0582-9. [DOI] [PubMed] [Google Scholar]
  6. van der Bliek A.M., Sedensky M.M., Morgan P.G. Cell biology of the mitochondrion. Genetics. 2017;207:843–871. doi: 10.1534/genetics.117.300262. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. de Boer R., Smith R.L., De Vos W.H., Manders E.M., Brul S., van der Spek H. Caenorhabditis elegans as a model system for studying drug induced mitochondrial toxicity. PLoS One. 2015;10:e0126220. doi: 10.1371/journal.pone.0126220. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Breckenridge D.G., Kang B.-H., Kokel D., Mitani S., Staehelin L.A., Xue D. Caenorhabditis elegans drp-1 and fis-2 regulate distinct cell-death execution pathways downstream of ced-3 and independent of ced-9. Mol. Cell. 2008;31:586–597. doi: 10.1016/j.molcel.2008.07.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Chan D.C. Mitochondrial dynamics and its involvement in disease. Annu. Rev. Pathol. Mech. Dis. 2020;15:235–259. doi: 10.1146/annurev-pathmechdis-012419-032711. [DOI] [PubMed] [Google Scholar]
  10. Chen M., Dai W., Sun S.Y., Jonasch D., He C.Y., Schmid M.F., Chiu W., Ludtke S.J. Convolutional neural networks for automated annotation of cellular cryo-electron tomograms. Nat. Methods. 2017;14:983–985. doi: 10.1038/nmeth.4405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Chen G., Chen P., Shi Y., Hsieh C.Y., Liao B., Zhang B. Rethinking the usage of batch normalization and dropout in the training of deep neural networks. arXiv. 2019 arxiv:1905.05928. [Google Scholar]
  12. Chlebus G., Schenk A., Moltz J.H., van Ginneken B., Hahn H.K., Meine H. Automatic liver tumor segmentation in CT with fully convolutional neural networks and object-based postprocessing. Sci. Rep. 2018;8:15497. doi: 10.1038/s41598-018-33860-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Falk T., Mai D., Bensch R., Cicek O., Abdulkadir A., Marrakchi Y., Bohm A., Deubner J., Jackel Z., Seiwald K. U-Net: deep learning for cell counting, detection, and morphometry. Nat. Methods. 2019;16:67–70. doi: 10.1038/s41592-018-0261-2. [DOI] [PubMed] [Google Scholar]
  14. De Fauw J., Ledsam J.R., Romera-Paredes B., Nikolov S., Tomasev N., Blackwell S., Askham H., Glorot X., O’Donoghue B., Visentin D. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018;24:1342–1350. doi: 10.1038/s41591-018-0107-6. [DOI] [PubMed] [Google Scholar]
  15. Di Fonzo A., Chien H.F., Socal M., Giraudo S., Tassorelli C., Iliceto G., Fabbrini G., Marconi R., Fincati E., Abbruzzese G. ATP13A2 missense mutations in juvenile parkinsonism and young onset Parkinson disease. Neurology. 2007;68:1557–1562. doi: 10.1212/01.wnl.0000260963.08711.08. [DOI] [PubMed] [Google Scholar]
  16. Grünewald A., Arns B., Seibler P., Rakovic A., Münchau A., Ramirez A., Sue C.M., Klein C. ATP13A2 mutations impair mitochondrial function in fibroblasts from patients with Kufor-Rakeb syndrome. Neurobiol. Aging. 2012;33:1843.e1–1843.e7. doi: 10.1016/j.neurobiolaging.2011.12.035. [DOI] [PubMed] [Google Scholar]
  17. Gusdon A.M., Zhu J., Van Houten B., Chu C.T. ATP13A2 regulates mitochondrial bioenergetics through macroautophagy. Neurobiol. Dis. 2012;45:962–972. doi: 10.1016/j.nbd.2011.12.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Ichishita R., Tanaka K., Sugiura Y., Sayano T., Mihara K., Oka T. An RNAi screen for mitochondrial proteins required to maintain the morphology of the organelle in Caenorhabditis elegans. J. Biochem. 2008;143:449–454. doi: 10.1093/jb/mvm245. [DOI] [PubMed] [Google Scholar]
  19. Ioffe S., Szegedy C. 2015. Batch normalization: accelerating deep network training by reducing internal covariate shift; pp. 448–456. (Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37). [Google Scholar]
  20. Kanazawa T., Zappaterra M.D., Hasegawa A., Wright A.P., Newman-Smith E.D., Buttle K.F., McDonald K., Mannella C.A., van der Bliek A.M. The C. elegans Opa1 homologue EAT-3 is essential for resistance to free radicals. PLoS Genet. 2008;4:e1000022. doi: 10.1371/journal.pgen.1000022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Kreshuk A., Zhang C. Machine learning: advanced image segmentation using ilastik. Methods Mol. Biol. 2019;2040:449–463. doi: 10.1007/978-1-4939-9686-5_21. [DOI] [PubMed] [Google Scholar]
  22. Krizhevsky A., Sutskever I., Hinton G.E. Curran Associates Inc.; 2012. ImageNet classification with deep convolutional neural networks; pp. 1097–1105. (Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 1). [Google Scholar]
  23. Kusumoto D., Yuasa S. The application of convolutional neural network to stem cell biology. Inflamm. Regen. 2019;39:14. doi: 10.1186/s41232-019-0103-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Labrousse A.M., Zappaterra M.D., Rube D.A., van der Bliek A.M. C. elegans dynamin-related protein DRP-1 controls severing of the mitochondrial outer membrane. Mol. Cell. 1999;4:815–826. doi: 10.1016/s1097-2765(00)80391-3. [DOI] [PubMed] [Google Scholar]
  25. Lambie E.J., Tieu P.J., Lebedeva N., Church D.L., Conradt B. CATP-6, a C. elegans ortholog of ATP13A2 PARK9, positively regulates GEM-1, an SLC16A transporter. PLoS One. 2013;8:e77202. doi: 10.1371/journal.pone.0077202. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. LeCun Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989;1:541–551. [Google Scholar]
  27. Li Y., Gong H., Wu W., Liu G., Chen G. An automated method using hessian matrix and random walks for retinal blood vessel segmentation. 2015 8th International Congress on Image and Signal Processing (CISP) 2015 [Google Scholar]
  28. Lindeberg T. Scale-space theory: a basic tool for analysing structures at different scales. J. Appl. Stat. 1994;21:225–270. [Google Scholar]
  29. Long J., Shelhamer E., Darrell T. Fully convolutional networks for semantic segmentation. arXiv. 2014 doi: 10.1109/TPAMI.2016.2572683. arxiv:1411.4038. [DOI] [PubMed] [Google Scholar]
  30. Marr D., Hildreth E. Theory of edge detection. Proc. R. Soc. Lond. B Biol. Sci. 1980;207:187–217. doi: 10.1098/rspb.1980.0020. [DOI] [PubMed] [Google Scholar]
  31. McQuin C., Goodman A., Chernyshev V., Kamentsky L., Cimini B.A., Karhohs K.W., Doan M., Ding L., Rafelski S.M., Thirstrup D. CellProfiler 3.0: next-generation image processing for biology. PLoS Biol. 2018;16:e2005970. doi: 10.1371/journal.pbio.2005970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Meijering E. Cell segmentation: 50 Years down the road [life sciences] IEEE Signal Process. Mag. 2012;29:140–145. [Google Scholar]
  33. Moller J.V., Juul B., le Maire M. Structural organization, ion transport, and energy transduction of P-type ATPases. Biochim. Biophys. Acta. 1996;1286:1–51. doi: 10.1016/0304-4157(95)00017-8. [DOI] [PubMed] [Google Scholar]
  34. Orozco-Fuentes S., Neganova I., Wadkin L.E., Baggaley A.W., Barrio R.A., Lako M., Shukurov A., Parker N.G. Quantification of the morphological characteristics of hESC colonies. Sci. Rep. 2019;9:17569. doi: 10.1038/s41598-019-53719-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Pernas L., Scorrano L. Mito-morphosis: mitochondrial fusion, fission, and cristae remodeling as key mediators of cellular function. Annu. Rev. Physiol. 2016;78:505–531. doi: 10.1146/annurev-physiol-021115-105011. [DOI] [PubMed] [Google Scholar]
  36. Ramirez A., Heimbach A., Grundemann J., Stiller B., Hampshire D., Cid L.P., Goebel I., Mubaidin A.F., Wriekat A.L., Roeper J. Hereditary parkinsonism with dementia is caused by mutations in ATP13A2, encoding a lysosomal type 5 P-type ATPase. Nat. Genet. 2006;38:1184–1191. doi: 10.1038/ng1884. [DOI] [PubMed] [Google Scholar]
  37. Ramonet D., Podhajska A., Stafa K., Sonnay S., Trancikova A., Tsika E., Pletnikova O., Troncoso J.C., Glauser L., Moore D.J. PARK9-associated ATP13A2 localizes to intracellular acidic vesicles and regulates cation homeostasis and neuronal integrity. Hum. Mol. Genet. 2012;21:1725–1743. doi: 10.1093/hmg/ddr606. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Regmi S.G., Rolland S.G. New imaging tools to analyze mitochondrial morphology in Caenorhabditis elegans. Methods Mol. Biol. 2017;1567:255–272. doi: 10.1007/978-1-4939-6824-4_16. [DOI] [PubMed] [Google Scholar]
  39. Rolland S.G., Lu Y., David C.N., Conradt B. The BCL-2-like protein CED-9 of C. elegans promotes FZO-1/Mfn1,2- and EAT-3/Opa1-dependent mitochondrial fusion. J. Cell Biol. 2009;186:525–540. doi: 10.1083/jcb.200905070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Rolland S.G., Motori E., Memar N., Hench J., Frank S., Winklhofer K.F., Conradt B. Impaired complex IV activity in response to loss of LRPPRC function can be compensated by mitochondrial hyperfusion. Proc. Natl. Acad. Sci. U S A. 2013;110:E2967–E2976. doi: 10.1073/pnas.1303872110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Ronneberger O., Fischer P., Brox T. Springer International Publishing; 2015. U-net: Convolutional Networks for Biomedical Image Segmentation, Cham. [Google Scholar]
  42. Sato Y., Nakajima S., Shiraga N., Atsumi H., Yoshida S., Koller T., Gerig G., Kikinis R. Three-dimensional multi-scale line filter for segmentation and visualization of curvilinear structures in medical images. Med. Image Anal. 1998;2:143–168. doi: 10.1016/s1361-8415(98)80009-1. [DOI] [PubMed] [Google Scholar]
  43. Scherer D., Müller A., Behnke S. Springer Berlin Heidelberg; 2010. Evaluation of Pooling Operations in Convolutional Architectures for Object Recognition. [Google Scholar]
  44. Schindelin J., Arganda-Carreras I., Frise E., Kaynig V., Longair M., Pietzsch T., Preibisch S., Rueden C., Saalfeld S., Schmid B. Fiji: an open-source platform for biological-image analysis. Nat. Methods. 2012;9:676–682. doi: 10.1038/nmeth.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Stember J.N., Chang P., Stember D.M., Liu M., Grinband J., Filippi C.G., Meyers P., Jambawalikar S. Convolutional neural networks for the detection and measurement of cerebral aneurysms on magnetic resonance angiography. J. Digit. Imaging. 2018;32:808–815. doi: 10.1007/s10278-018-0162-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Szegedy C., Liu W., Jia Y., Sermanet P., Reed S.E., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A. Going deeper with convolutions. arXiv. 2014 arxiv:1409.4842. [Google Scholar]
  47. Taha A.A., Hanbury A. Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC Med. Imaging. 2015;15:29. doi: 10.1186/s12880-015-0068-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Tan F.J., Husain M., Manlandro C.M., Koppenol M., Fire A.Z., Hill R.B. CED-9 and mitochondrial homeostasis in C. elegans muscle. J. Cell Sci. 2008;121:3373–3382. doi: 10.1242/jcs.032904. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Tilokani L., Nagashima S., Paupe V., Prudent J. Mitochondrial dynamics: overview of molecular mechanisms. Essays Biochem. 2018;62:341–360. doi: 10.1042/EBC20170104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Tondera D., Grandemange S., Jourdain A., Karbowski M., Mattenberger Y., Herzig S., Da Cruz S., Clerc P., Raschke I., Merkwirth C. SLP-2 is required for stress-induced mitochondrial hyperfusion. EMBO J. 2009;28:1589–1600. doi: 10.1038/emboj.2009.89. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Torborg C.L., Feller M.B. Unbiased analysis of bulk axonal segregation patterns. J. Neurosci. Methods. 2004;135:17–26. doi: 10.1016/j.jneumeth.2003.11.019. [DOI] [PubMed] [Google Scholar]
  52. van Veen S., Martin S., Van den Haute C., Benoy V., Lyons J., Vanhoutte R., Kahler J.P., Decuypere J.-P., Gelders G., Lambie E. ATP13A2 deficiency disrupts lysosomal polyamine export. Nature. 2020;578:419–424. doi: 10.1038/s41586-020-1968-7. [DOI] [PubMed] [Google Scholar]
  53. Wai T., Langer T. Mitochondrial dynamics and metabolic regulation. Trends Endocrinol. Metab. 2016;27:105–117. doi: 10.1016/j.tem.2015.12.001. [DOI] [PubMed] [Google Scholar]
  54. Wiesel T.N., Hubel D.H. Single-cell responses in striate cortex of kittens deprived of vision in one eye. J. Neurophysiol. 1963;26:1003–1017. doi: 10.1152/jn.1963.26.6.1003. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Document S1. Transparent Methods and Figures S1–S5
mmc1.pdf (1.5MB, pdf)

Data Availability Statement

The software documentation for the MitoS and MitoA tool can be found at https://github.com/mitosegnet. The MitoSegNet segmentation model, the MitoA analysis and MitoS segmentation tool (GPU/CPU) for Linux and Windows are available at https://zenodo.org/search?page=1&size=20&q=mitosegnet. The Python code used for generating the figures displayed in the manuscript is available at https://github.com/MitoSegNet/MitoSegNet_AccuracyTesting_Manuscript.


Articles from iScience are provided here courtesy of Elsevier

RESOURCES