Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2022 Mar 14;18(3):e1009949. doi: 10.1371/journal.pcbi.1009949

Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images

Thierry Pécot 1,*, Maria C Cuitiño 2, Roger H Johnson 3, Cynthia Timmers 4,¤, Gustavo Leone 5
Editor: Attila Csikász-Nagy6
PMCID: PMC8947602  PMID: 35286300

Abstract

Automatic characterization of fluorescent labeling in intact mammalian tissues remains a challenge due to the lack of quantifying techniques capable of segregating densely packed nuclei and intricate tissue patterns. Here, we describe a powerful deep learning-based approach that couples remarkably precise nuclear segmentation with quantitation of fluorescent labeling intensity within segmented nuclei, and then apply it to the analysis of cell cycle dependent protein concentration in mouse tissues using 2D fluorescent still images. First, several existing deep learning-based methods were evaluated to accurately segment nuclei using different imaging modalities with a small training dataset. Next, we developed a deep learning-based approach to identify and measure fluorescent labels within segmented nuclei, and created an ImageJ plugin to allow for efficient manual correction of nuclear segmentation and label identification. Lastly, using fluorescence intensity as a readout for protein concentration, a three-step global estimation method was applied to the characterization of the cell cycle dependent expression of E2F proteins in the developing mouse intestine.

Author summary

Estimating the evolution of protein concentration over the cell cycle is an important step towards a better understanding of this key biological process. Unfortunately, experimental designs to monitor proteins in individual living cells are expensive and difficult to set up. We propose instead to consider 2D images from tissue biopsies as snapshots of cell populations to reconstruct the actual protein concentration evolution over the cell cycle. This requires to accurately localize cell nuclei and identify nuclear fluorescent proteins. We take advantage of the powerful deep learning technology, a machine learning approach which has revolutionized computer vision, to achieve these challenging tasks. Additionally, we have created an ImageJ plugin to quickly and efficiently annotate images or correct annotations, required to build training datasets to feed the deep convolutional neural networks.


This is a PLOS Computational Biology Methods paper.

Introduction

Automatic image analysis is at the core of human and animal tissue-based research. However, quantitation of morphological features or fluorescent labeling in intact mammalian tissues still remains a challenge. The densely packed nuclear aggregates that characterize many of these tissues, the extensive variability across different tissue types, and the continuously increasing number of imaging modalities are some of the many variables that make tissue biological quantification an extremely difficult task. Over the last decade [14], deep learning has brought artificial intelligence to the forefront of image-based decision making. In particular, deep convolutional neural networks have demonstrated their superiority for image segmentation [1, 2]. These approaches have also outperformed the traditional approaches used in microscopy, such as watershed for nuclei or cell segmentation [59]. However, this machine learning-based approach requires large amounts of annotated data and new strategies have to be developed to process highly complex biological objects acquired with different modalities by considering small training datasets. In this paper, we propose a series of deep learning-based approaches to precisely segment nuclei and to identify fluorescently labelled cells in order to analyze the evolution of cell cycle dependent E2F protein concentration in mouse tissues.

E2Fs are major regulators of the cell cycle. The members of this family of transcription factors are categorized into three subclasses in mammals: canonical activators (E2F1–3A/B), canonical repressors (E2F4–6), and atypical repressors (E2F7–8) [1013]. Adding to the body of literature on E2F-dependent transcriptional activity in vivo, our lab previously provided quantitative evidence on the temporal expression of representative activator (E2F3A), canonical repressor (E2F4) and atypical repressor (E2F8) family members during embryonic development [14], all of which have been shown to be of major importance [1518]. To establish the temporal expression profiles of the three sentinel E2Fs, we used an E2F3A specific antibody and generated MYC-tagged E2F4/E2F8 knock-in mice. In addition to fluorescence labeling of E2F3A, E2F4 and E2F8, 5-ethynyl-2’-deoxyuridine (EdU) and Histone H3 S10 phosphorylation (pH3) were used to identify S, G2 and M phases. Images of eight different combinations of markers were acquired from sections of the developing mouse intestine using confocal and widefield microscopy (see S1 Fig). The data analysis pipeline consisted of i) nuclear segmentation with a deep learning approach [19], ii) nuclear marker identification by thresholding, and iii) estimation of E2F concentrations over the cell cycle from 2D intensity histograms.

In this manuscript, we propose and evaluate alternative methods to quantify nuclear protein levels which result in an improved automated pipeline with greatly reduced requirement for interactive manual corrections. Using a small training dataset composed of 2D still images (with and without various forms of data augmentation), we first evaluate five different deep learning strategies to segment nuclei in microscopic images of embryonic mouse intestinal epithelium. We also design post-processing methods to improve nuclear segmentation. We then propose another deep learning-based approach for identifying nuclear markers in the epithelial cells and demonstrate the superiority of this method to the usual threshold based method [20, 21]. Additionally, we create an ImageJ plugin [22, 23] named Annotater to specifically and efficiently correct nuclear segmentation and marker identification, ensuring that nuclear features are accurately quantified. Next, these image features extracted from 2D still images are used to perform a temporal analysis of E2F protein concentration over the cell cycle. Based on three mathematical assumptions grounded in cell biology, we initialize the temporal evolution of E2F concentrations using a graph optimization method. Cell cycle markers are then used to temporally register E2F proteins’ concentration with respect to cell cycle phase. The global estimation of the protein concentration of E2F3A, E2F4 and E2F8 through the cell cycle is defined as an assignment problem and solved with the Hungarian algorithm [24, 25]. This approach is extensively evaluated with simulated data. Finally, we directly estimate the temporal evolution of E2F concentrations without using marker identification. In addition, we evaluate the impact of using different amount of images in the training datasets for nuclei segmentation on the estimation of E2F concentrations over the cell cycle.

Results

Mask R-CNN is the optimal deep learning approach for segmenting nuclei in cross-sectional images of complex tissues

Over the last decade, deep learning has revolutionized computer vision [14]. Over the years, several deep learning approaches have been successfully applied to segment cell nuclei [19, 26, 27]. More recently, the 2018 data science bowl [28] attempted to definitely solve this problem in 2D by posing the challenge: Create an algorithm for automated nuclei detection. Although impressive results were obtained for different light microscopy modalities and a variety of nuclear stains [7], nuclear segmentation from complex tissues such as intestinal epithelium still represents an unusually challenging problem. As is the case for most biological imaging studies due to natural variability in the objects of interest that are captured using dissimilar imaging modalities, no annotated data is available for this specific application. Consequently, we set out with the goal of designing a robust approach that would lend itself to routine use in the typical biology laboratory. We evaluated five different deep learning approaches for nuclear segmentation: U-Net [27, 29], Inception-V3 [30], Mask R-CNN [31], Stardist [8] and CellPose [9]. U-Net, perhaps the most-used deep convolutional neural network for biomedical data, is composed of an encoder part, used to capture image features, and a decoder part to estimate a class at each pixel of the input images. Inception-V3, which was designed to identify objects in images, is a deep convolutional neural network that only estimates one class given an input image. When using this architecture for nuclear segmentation, inputs are defined as image patches and the output corresponds to the class at the patch center. Inception-V3 is much slower than U-Net for both training and processing (S1 Table), as a decision is only made for the central pixel of the input image patch, in contrast to a decision being made at each pixel of the input image as it is in U-Net. For both U-Net and Inception-V3, three classes are defined: inner nuclei, nuclei contours, and background [19, 32]. Individual nuclei are then obtained by subtracting the nuclei contours from the inner nuclei (see Materials and methods). To improve performance, we developed a post-processing method that we call corrected watershed, wherein the results obtained with the U-Net or Inception-V3 network are combined with those produced by the watershed method [5] (see Materials and methods). Mask R-CNN, Stardist and Cellpose are instance segmentation approaches, i.e. they directly estimate individual objects. These three methods first extract image features using a backbone convolutional neural network. Stardist identifies individual objects by predicting the distances to object boundaries with a fixed set of rays, ending up with a set of polygons for a given input image, corresponding to nuclei. Cellpose predicts an alternative representation of the nuclei masks, the equilibrium distribution of a heat-diffusion simulation with a heat source placed at the center of the mask. Mask R-CNN relies on a Region Proposal Network (RPN) to submit subregions of the input image. A fully connected neural network then defines a class and a bounding box for the input subregions, and a convolutional neural network generates a segmentation mask for the same input subregions. With these three approaches, the output corresponds directly to individual nuclei. The five approaches and their use are described in more details in the Materials and methods. Pre-processing, normalization, optimization, data augmentation, transfer learning and post-processing are summarized for each method in S2 Table. The number of images and nuclei used in the training and validation datasets are shown in S3 Table. Two different modalities, confocal and widefield images are used in this study, which allows us to test the genericity of the deep learning approaches for nuclei segmentation.

As proposed by Caicedo et al. [32], we use the F1 score with respect to the Intersection over Union (IoU) to compare performance of the three deep learning approaches (see Materials and methods). There are two aspects to performance: 1) the ability to correctly identify all nuclei in an image, which is the same as detecting the correct number of nuclei, and 2) the accuracy of the nuclear contours created in the output. This second aspect of performance is quantified by the F1 score over a range of IoU values. The IoU of two objects is the ratio of the intersection to the union of their areas. With an IoU of at least 0.5 (within the range 0.5–1.0), only two nuclei, one from the ground truth as determined by a pathologist and one estimated with a given method, can be paired. The F1 score for an IoU equal to 0.5 can be used to assess the ability of a method to accurately identify all nuclei. F1 scores for IoU thresholds in the range 0.5–1.0 reflect the localization precision of the segmentation, which means the accuracy of the defined nuclear contours.

The three instance segmentation approaches, i.e. Stardist, Cellpose and Mask R-CNN, show similar performance (see S4S6 Figs and S5 and S6 Tables), with a lower accuracy for Cellpose. This approach has been designed to segment cells with cytoplasmic markers and is therefore less efficient with nuclear markers. Stardist shows a slightly better F1 score for an IoU threshold equal to 0.5 for confocal images than Mask R-CNN (0.864 vs 0.858), suggesting that a few more nuclei are identified with Stardist. However, the F1 scores for higher thresholds of IoU obtained with Mask R-CNN are clearly higher than those obtained with Stardist (IoU = 0.75, 0.458 vs 0.434 for confocal images and 0.460 vs 0.389 for widefield images). This demonstrates that Mask R-CNN better localizes the nuclei areas than Stardist. Of note, while pooling together confocal and widefield images leads to a better accuracy for all three methods, data augmentation does not improve performance for Stardist and Cellpose while it clearly ameliorates Mask R-CNN precision. This could suggest that Stardist and Cellpose are more sensitive to noise when the training dataset is small and the nuclei are densely packed.

In the following, only Mask R-CNN is compared to U-Net and Inception-V3. Fig 1 shows the superiority of the Mask R-CNN method for both modalities. The results represent the best performance of each deep learning method from among the algorithmic variants that invoke combinations of transfer learning, data augmentation and watershed correction as discussed below in connection with S2S4 Figs. S5 and S6 Tables show the F1 score obtained with IoU = 0.5 and IoU = 0.75 when considering all algorithmic variants for each method.

Fig 1. Comparison of U-Net, Inception-V3 and Mask R-CNN for nuclear segmentation.

Fig 1

a-b F1 score for range of IoU thresholds obtained with the U-Net, Inception-V3 and Mask R-CNN approaches for confocal a and widefield b images. Lines correspond to average F1 score over the three tested images while the shaded areas represent the standard error. c-d Segmented nuclei obtained with the U-Net (blue), Inception-V3 (green) and Mask R-CNN (cyan) as well as the ground truth (red) for confocal c and widefield d images. White inset rectangles in top images are shown at higher magnification below for the ground truth and each approach. In high-mag images, correctly identified nuclei are outlined in red, incorrectly in green. Scale bar = 20μm.

As shown in Fig 1a and 1b, and emphasized in the magnified images of Fig 1c and 1d, the U-Net method produced the worst results. Data augmentation, a process that artificially and massively increases the size of the training dataset by applying mathematical operations such as adding noise and rotating or flipping the images (see Materials and methods), clearly improves the segmentation accuracy (S2c and S2d Fig). However, pooling together confocal and widefield images in the training dataset produces worse segmentation accuracy compared to training on only confocal or only on widefield images. This demonstrates the inability of the U-Net approach to generalize nuclear segmentation with a small training dataset.

In contrast, the Inception-V3 approach obtains better results when both modalities are used in the training dataset (S3 Fig). While for Inception-V3 the training dataset was not increased when applying data augmentation because the computation time is already long, the high degree of overlap between input patches to which mathematical operations are applied (see Materials and methods) produces an effect similar to data augmentation, explaining in part the improved performance of the Inception-V3 approach relative to U-Net. Because Inception-V3 estimates the class one pixel at a time, both training and processing are highly compute intensive and not feasible in practice (S3d Fig). The proposed corrected watershed post-processing improves the results obtained for both U-Net and Inception-V3 (S2e, S2f, S3e and S3f Fig and S5 and S6 Tables).

The Mask R-CNN method obtains the best nuclear segmentation for both modalities (Fig 1), in large part due to the improvement realized through data augmentation and in a lesser extent to the transfer learning from the coco dataset [33], (Fig 1a and 1b). Combining confocal and widefield images in the training dataset improves the results, especially for confocal images (S4c–S4e Fig). This demonstrates how Mask R-CNN can benefit from an increase in training dataset size (from 4847/1619 nuclei in the training/validation dataset for confocal images alone to 15541/4051 nuclei in the training/validation dataset when confocal and widefield images are pooled together) even if the data are not uniform, coming from different sources. Of note, combined with massive data augmentation, transfer learning from the coco dataset really improves accuracy when considering confocal images only (S4e Fig compared to S4c Fig) or widefield images only (S4f Fig compared to S4d Fig), while it does not drastically improve the accuracy when confocal and widefield images are pooled together (S4c–S4f Fig). This suggests than considering images from different modalities has a similar effect for convergence to a plateau with Mask R-CNN than transfer learning with a pre-trained model. Moreover, training and processing are fast (S4d Fig), and the results do not require any post-processing. One limitation of the Mask R-CNN method compared with U-Net and Inception-V3 is an inferior boundary localization accuracy for confocal images, as demonstrated by the lower F1 scores for IoU threshold values greater than 0.75 (Fig 1a). Due to slightly larger nuclear masks (cyan boundaries in Fig 1c), this limitation is more than compensated for by Mask R-CNN’s higher true positive nuclei identification rate in most biological applications. While the performance of Mask R-CNN is impressive, its accuracy is not perfect and may be insufficient for many applications. Therefore, we designed the ImageJ plugin Annotater (see Materials and methods), a tool that allows users to efficiently correct the nuclear segmentation.

Deep learning improves identification of fluorescent nuclear markers

After completion of nuclei segmentation with the DAPI channel, it is possible to use the fluorescence in the other channels to extract information of interest about the cells. E2Fs positive/negative status can be evaluated that way, as well as EdU and pH3 patterns. EdU and pH3 are cell cycle markers that show evolving patterns along the cell cycle [14]. EdU is diffuse during first half of S phase and becomes punctate during second half of S phase. pH3 is first punctate during second half of S phase and G2, and becomes diffuse during mitosis. Typically, a thresholding procedure is applied to identify nuclear markers [20, 21], but this approach is not always accurate, especially when different patterns of fluorescence over a wide range of intensities are involved as is the case for EdU and pH3 (see diffuse and punctate patterns in Fig 2). To improve accuracy, we tested a deep learning approach for nuclear marker identification. As the goal is not to identify regions but to make a decision for each nucleus regarding the presence/absence of an E2F or the diffuse/punctate/absence of EdU or pH3, the instance segmentation and U-Net approaches are not suitable. In contrast, with an input defined as an image patch centered on each nucleus, the Inception-V3 architecture is appropriate to decide about the presence, potentially in diffuse or punctate state, or absence of a marker. In addition to data augmentation, we also define a so-called pixel-based training dataset (as opposed to nuclei-based training dataset) that includes the input patches centered at each pixel belonging to the nuclei (see Materials and methods). This strategy has a similar effect to data augmentation, as it drastically increases the training dataset. Although DAPI staining is different from the nuclear markers used to identify the E2Fs, EdU and pH3, the images are acquired simultaneously, so the image features captured by the Inception-V3 method for nuclear segmentation are potentially meaningful to identify nuclear markers. Consequently, we also perform transfer learning from the nuclear segmentation (see Materials and methods). To easily set the threshold for marker identification, we designed an interface in the Annotater that we used to obtain the results shown in Fig 2 for manual thresholding.

Fig 2. Comparison between Inception-V3 and manual thresholding for marker identification.

Fig 2

a-b Top rows (images): examples of E2F3A-, E2F8- and E2F4-positive and -negative nuclei, as well as EdU- and pH3-negative, diffuse and punctate nuclei in a confocal and b widefield images. Bottom rows (bar graphs): Accuracy obtained with the Inception-V3 and manual thresholding approaches for marker identification of E2F3A, E2F8, E2F4, EdU and pH3 in a confocal and b widefield images.

As shown in Fig 2, compared to manual thresholding, the Inception-V3 approach provides better performance for each marker in both modalities. Manual thresholding achieves a relatively good performance for E2F3A and E2F8 identification in both modalities, and for pH3 in confocal images. The latter might appear surprising, but the two different patterns for pH3 in confocal images are different enough to allow a strategy based on the thresholded area in the nuclei (see Materials and methods) to lead to satisfying results (Fig 2a). However, the results for E2F4, Fig 2b, are not as good: Because E2F4 is also cytoplasmic, the extra-nuclear fluorescence confounds the thresholding decision. Finally, thresholding clearly fails to identify the two different patterns (diffuse, punctate) of EdU in images from both modalities as well as the patterns of pH3 in widefield images. In contrast, the Inception-V3 approach yields accuracies greater than 90% for all markers except E2F4 (89%) (Fig 2). As shown in S7 Table, the use of the pixel-based training datasets does not significantly improve marker identification for confocal images, but does improve performance on E2F3A, E2F4 and pH3 markers in widefield images (see S8 Table). Additionally, the transfer learning from the nuclear segmentation slightly improves the results for all markers in both modalities. Computation (after training) is fast because only one decision is made per nucleus (S9 Table). Overall, this study demonstrates the remarkable accuracy of the deep learning-based approach in identifying cells that are positive for the tested nuclear markers, not only when the marker is both nuclear and cytoplasmic, but also when it exhibits different labeling patterns, such as diffuse and punctate. Given such a high level of accuracy, correcting the results with the Annotater plugin takes only a short time.

Estimation of E2F accumulation over the cell cycle from 2D still images

Estimating cell cycle progression is an important biological question and can be answered by using images [34, 35]. Unfortunately, these machine learning approaches require annotated data to be processed. Without training data, the most straightforward way to assess the evolution of a protein’s concentration over the cell cycle would be to monitor the expression level in every cell through all phases of the cycle. However, this is not possible in the proliferative tissue of a living animal. Instead of observing one cell, we propose to observe a large population of cells, each corresponding to a snapshot characterizing the cell state at a particular time during the cell cycle, and to reconstruct protein concentration as a function of time in the cell cycle by combining all these snapshots. As fluorescence intensity is proportional to protein concentration [36, 37], we define quantized levels of intensity for E2Fs (see S7 Fig and Materials and methods). In each individual image, we measure the average fluorescence intensity for each positive cell. Within each image, the range of intensity from the lowest average intensity to the highest is divided into bins that define the levels of intensity. On the other hand, EdU and pH3 markers are defined by their diffuse and punctate states, which change as a function of cell cycle phase. As shown in our previous work [14], EdU shows a diffuse pattern during the first half of S phase and a punctate pattern during the second half of S phase. pH3 shows a punctate pattern during the second half of S phase and G2 and a diffuse pattern during mitosis. These two markers allow us to register E2Fs expression with respect to the cell cycle. We propose to globally estimate the E2F concentration’s evolution over the cell cycle in three steps: i) initialization, ii) cell cycle registration and iii) global optimization. In this paper, as in our previous manuscript [14], the phrase “protein accumulation” is synonymous with “protein concentration” and connotes this balance between production and degradation, regardless of whether the protein concentration is increasing or decreasing. Also in this paper, the fluctuation over time of a quantity such as fluorescence intensity or protein accumulation is referred to as “evolution,” which is therefore synonymous with “time course.”

We use the term “initialize” to mean creation of a first estimate of protein accumulation over the cell cycle: a graph of fluorescence intensity vs. time. To initialize all combinations of markers from the images (see S1 Fig), we make three fundamental assumptions:

  1. The number of cells in a given phase of the cell cycle is proportional to the duration of that phase.

  2. Temporal evolution of protein accumulation is similar in all observed cells (there are no subpopulations with different cell cycle evolution).

  3. Concentrations of E2Fs evolution can be represented by concave downward parabolas, i.e. they increase from 0 to their maximum and then decrease from this maximum to 0.

These assumptions are validated by the following biological statements:

  1. Proliferation in intestinal epithelium is asynchronous, so each cell cycle phase is observed at a frequency proportional to its duration.

  2. All cells in intestinal epithelium are proliferating and undergo cell cycle at a uniform rate.

  3. Cell cycle-regulated proteins first accumulate over time and are then degraded.

We use 2D histograms to initialize protein accumulation (S8 Fig and Materials and methods). In these histograms, the first axis corresponds to the intensity of an E2F, while the second axis corresponds to either another E2F intensity or the states of EdU or pH3 staining. From these histograms, we define graphs with costs associated with edges reflecting the possible intensity/pattern transitions of the E2Fs, EdU and pH3. The sequence of edges that goes through all vertices with minimum cost is used to initialize the evolution of protein accumulation over time (S9 Fig and Materials and methods). The initializations obtained for all eight different combinations of markers are shown in S10 Fig. As EdU and pH3 labeling patterns with respect to the cell cycle are known, we then register E2Fs with respect to EdU and pH3 through a circular permutation (see Materials and methods). We use the same approach to refine EdU labeling during the cell cycle (see Materials and methods). The protein accumulations after registration are shown in S11 Fig. We finally model the global estimation of E2F accumulation over the cell cycle as an assignment problem. We use the Hungarian algorithm [24, 25] to successively estimate the individual accumulation of each E2F over the cell cycle given the EdU and pH3 patterns as constraints (see Materials and methods). We also add a local constraint to only allow growing substitutions for E2Fs (see Materials and methods). The final result is shown in Fig 3, depicting the successive waves of E2F3A, E2F4 and E2F8 over the cell cycle.

Fig 3. Estimation of E2Fs evolution over the cell cycle.

Fig 3

Temporal evolution of E2F3A, E2F8 and E2F4 protein accumulation, as well as EdU and pH3 patterns over the cell cycle in mouse intestinal epithelium.

We then simulate data to evaluate the validity of our approach (see Materials and methods). In summary, E2F3A and E2F8 are randomly generated to have a concentration that shows a concave downward parabola representation while E2F4 is randomly generated to have a concentration that shows two successive concave downward parabola representations, with random durations for each intensity level for all three E2Fs. EdU and pH3 are also randomly generated to show a succession of punctate/diffuse or diffuse/punctate patterns. Noise is randomly added to corrupt from 0% to 50% of temporal bins. For each set of parameters, 5 different simulations are randomly generated. The performance is evaluated by measuring the mean squared error (MSE) between the generated and estimated concentrations of E2Fs, EdU and pH3. A first set of simulations estimates the influence of the number of samples, corresponding to the number of mice, and shows that the MSE drastically decreases from 1 sample to 2 samples (see S12 Fig), suggesting that 3 mice for E2F3A, E2F8, EdU and pH3, and 2 mice for E2F4 should lead to a satisfying estimation. A second set of simulations evaluates the influence of the number of bins used for time. As shown in S13 Fig, the MSE is low for a range of 20 to 100 bins and drastically increases for 125 bins, validating the choice of 100 bins for time. A last set of simulations estimates the impact of the number of bins used for intensity. The MSE is quite stable for a range of 2 to 5 bins but ramps up with 10 bins (see S14 Fig). This confirms our choice to use 4 bins, which is also widely used in both diagnostic pathology and biomedical research [38, 39]. Fig 4 summarizes the evaluation with simulations when considering the same parameters (number of samples, intensity and time bins) than those used for real data with up to 50% of the simulated data corrupted with noise.

Fig 4. Evaluation of the estimation of protein concentration over the cell cycle on simulated data.

Fig 4

Mean squared error between the estimated and simulated concentrations of E2F3A, E2F8, E2F4, EdU and pH3 when corrupting up to 50% of the simulated data with noise and considering 3 samples for E2F3A, E2F8, EdU and pH3, 2 samples for E2F4, 100 bins for time and 4 bins for intensity. The lines correspond to the average mean squared error while the areas represent the standard error.

Estimation of E2F accumulation over the cell cycle with respect to nuclei segmentation

In the previous section, we used the marker identification to identify negative cells for E2Fs and the two patterns for EdU and pH3. We now want to assess if this step is required or not. Fig 5a shows the estimation of E2Fs concentration when considering only positive/negative states for EdU and pH3. While the estimated concentration slightly changes, especially the second peak of E2F4, the successive waves of E2Fs are preserved in time. This opens the possibility to a more direct concentration estimation where a nuclei segmentation is performed first and the intensity is then binned into four levels for E2Fs channels and two levels for EdU and pH3 (see Materials and methods). Fig 5b depicts the estimated concentrations for E2Fs with a manual segmentation for nuclei. The estimated concentrations are more degraded than with a better estimation of EdU and pH3 states, but E2Fs still show 4 successive peaks which are temporally well estimated, with the exception of the first E2F4 peak which is slightly delayed. This approach allows to evaluate the impact of nuclei segmentation on the concentration estimation over the cell cycle. We estimate the E2Fs concentration with the nuclei segmentation obtained for the confocal images with a Mask R-CNN model trained with the widefield images and a manual segmentation for the widefield images (see Fig 5c), and conversely (see Fig 5d). These estimations are actually close to the one obtained with manual segmentation, i.e. a degraded accuracy as shown in Fig 5e, but a well estimated temporality for the successive waves of E2Fs concentrations. This illustrates the robustness of our approach with respect to nuclei segmentation to temporally estimate the waves of E2Fs over the cell cycle.

Fig 5. Estimation of E2Fs accumulation over the cell cycle without marker identification with respect to nuclei segmentation.

Fig 5

Temporal evolution of E2F3A, E2F8 and E2F4 protein accumulation over the cell cycle in mouse intestinal epithelium, by considering positive/negative states for EdU and pH3 a, by only considering intensity for E2Fs, EdU and pH3 with a manual segmentation for nuclei b, with a nuclei segmentation obtained for the confocal images with a Mask R-CNN model trained with the widefield images and a manual segmentation for the widefield images c, and with a nuclei segmentation obtained for the widefield images with a Mask R-CNN model trained with the confocal images and a manual segmentation for the confocal images d.

Discussion

This study demonstrates that the Mask R-CNN approach, coupled with transfer learning and data augmentation, can produce highly accurate nuclear segmentation of fluorescent images, even in a complex tissue such as the intestinal epithelium, by considering a small training dataset. Outstanding segmentation results are achieved for different imaging modalities, notably including widefield fluorescence microscopy, a ubiquitous imaging mode in cell biology research labs that produces challenging images of low sharpness and poor boundary definition. This is of major interest, as segmentation is a key task in image analysis of mammalian tissues where variability across tissue types is high and the number of modalities large and ever increasing. Therefore, the quest for comprehensive training datasets will be long and biology laboratories need efficient strategies to process their own images. In this context, the Inception-V3 approach also obtains satisfactory segmentation accuracy, but the computation time, especially for training, represents a major limitation. Though the U-Net approach has demonstrated its utility in many biomedical studies, it is unable to generalize the ability to segment nuclei when the images are acquired with different modalities.

Our results also attest to the superiority and efficiency of the deep learning approach for identifying the presence and state of nuclear markers. As the goal is to make a decision for each cell, the Inception-V3 architecture is well adapted to the task, computation time not being the issue that it is for nuclear segmentation. Considering pixel-based input patches improves performance if the training dataset is small. Moreover, transfer learning from nuclear segmentation improves marker identification. This is explained by the fact that image features that are suited for nuclear segmentation are also meaningful for identifying nuclear markers. This is an interesting observation, as the same strategy could be used in an n-steps pipeline analysis. Indeed, a deep learning classifier could be trained for the first analysis applied to a particular dataset, the resulting parameters used to initialize the classifier for the second analysis, and so on.

While deep learning approaches for both nuclear segmentation and nuclear marker identification show high accuracy, some level of discrepancy between desired and actual results is expected. Thus, it is important to allow the user to correct the results if higher accuracy is required. This need for interactive post-processing of the results motivated us to design and implement the ImageJ plugin Annotater. This plugin can be used to annotate data from scratch or to perform corrections. While Annotater could be used to obtain perfect accuracy even without prior automated segmentation and marker identification, starting from a good estimate markedly speeds up the process. For instance, delineating the boundaries of all nuclei in an image of the intestinal epithelium may take a trained user 1–1.5 hours whereas any user can correct the results obtained with the Mask R-CNN approach in less than five minutes.

We present a novel and powerful method to estimate temporal phenomena using quantitative analysis of 2D still images. Specifically, we showed how the time-variant concentrations of intra-nuclear proteins can be qualitatively obtained from snapshot images of cell populations of reasonable size (for example, FFPE images of hundreds of epithelial cells as we analyzed in this study). Compared to alternative methods like single-cell analysis of individual living cells, the method we put forward here potentially allows easier and less expensive experimental designs for many applications in which dynamic protein expression information is required. In this particular case, a smart initialization based on biologically relevant assumptions coupled with a global optimization approach enabled estimation of E2F accumulation over the cell cycle. We also demonstrate that this approach is robust to nuclei segmentation and does not require a marker identification to provide a good temporal estimation of the successive E2Fs waves of concentration.

Materials and methods

Mouse models

The E2f4myc/myc and E2f8myc/myc were generated using standard homologous recombination cloning techniques as described in [14]. E2f4myc/myc, E2F8myc/myc and wild type controls were maintained on a mixed background (FVB/NT, 129v/Sv, C57BL/6NT). Tissues were collected from E13.5 intestinal epithelium.

Ethics statement

Mouse usage and protocols were approved by the Institutional Animal Care and Use Committee at the Ohio State University and Medical University of South Carolina. Mice were housed under normal husbandry conditions (five or less animals per cage) in a vivarium with a 12-hour light/dark cycle.

Immunostaining

Immunostaining was performed on a Bond Rx (Leica) or Ventana Discovery ULTRA (Roche) autostainer as per manufacturer’s instructions as previously described [40, 41]. Primary antibodies and dilutions used in this study were as follows: pH3-S10 (Millipore; 06–570, 1:250), E2f3a (Millipore; 05–551, 1:100) and Myc-tag (Cell Signaling Technology; 2278, 1:100). EdU staining was performed following the manufacturer’s protocol (Life Technologies; C10337).

Image acquisition

Widefield micrographs were collected using a Nikon Eclipse Ni-U microscope with a DS-Qi2 camera and NIS-Elements Advanced Research software. Confocal micrographs were collected using the Olympus FV 1000 Filter Confocal system in the Campus Microscopy and Imaging Facility at the Ohio State University.

Nuclei segmentation

The five deep learning approaches were coded in Python with the Python libraries numpy [42], tensorflow [43], PyTorch [44], keras [45], scipy [46] and scikit-image [47]. The code with the parameters used to train and process all experiments presented in this manuscript is available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/. It was written with jupyter notebooks and widgets in order to be used by biologists. Video tutorials are also available on the same web page.

Training dataset

The training dataset for nuclei segmentation consisted of twelve confocal and forty-five widefield 512 x 512 images annotated by a moderately-skilled researcher. Four different training datasets were used to evaluate the deep learning approaches and are summarized in S3 Table For the Inception-V3 approach, training and validation datasets were pulled together to obtain the input patches. Ninety percent of image patches were used for training and ten percent of image patches were used for validation.

U-Net

As the U-Net approach estimates a class at each pixel, three classes were defined to allow separating nuclei as proposed in [19]: inner nuclei, nuclei contours and background. To facilitate nuclei separation, the nuclei contours in the training dataset were dilated [19]. To reduce memory usage and limit over-fitting, the imaging field for images in the training dataset was set to 256 x 256 by randomly cropping the 512 x 512 input images. These cropped images were then normalized with a 1–99 quantile. A root mean square prop was used to estimate the parameters of the deep convolutional neural network by minimizing a weighted cross entropy loss to handle class imbalance for 100 epochs. The weights associated with each class were defined from the training dataset as their inverse proportion. A data augmentation to increase the training dataset by a factor of 100 was processed before normalization with the imgaug python library [48] and included flipping, rotation, shearing, blurring, sharpness and brightness modifications, noise addition, intensity inversion and contrast modifications.

Inception-V3

The same three classes were defined and the nuclei contours were dilated for the Inception-V3 approach as for the U-Net approach. The imaging field was set to 65 x 65 pixel image patches, large enough to include at least one nucleus as suggested in [19]. For each image patch, only the class of the central pixel was trained or processed. The training images were normalized by dividing the intensity at each pixel by the median intensity over the image and by subtracting at each pixel the average intensity of a 65 x 65 neighborhood around it. To obtain a balanced training dataset, the same number of input image patches were defined for each class. A stochastic gradient descent was used to estimate the parameters of the deep convolutional neural network by minimizing a categorical cross entropy for 15 epochs. A modification of the input image patches was processed after normalization with the imgaug python library [48] and included the same image transformations that were used for the U-Net approach.

U-Net and Inception-V3 post-processing

For both U-Net and Inception-V3 approaches, nuclei were obtained by subtracting the nuclei contours from the inner nuclei. Let us define Sc = {Sc(x)}x∈Ω where Sc(x) denotes the nuclei contour score obtained with the U-Net or the Inception-V3 approach at pixel x, and Ω is the regular grid of pixels in the image. Similarly, we define Sn = {Sn(x)}x∈Ω the inner nuclei score at each pixel x. The nuclei component N = {N(x)}x∈Ω is defined as:

N(x)=Sn(x)-Sc(x). (1)

This nuclei component is then thresholded to define a binary image Nb:

Nb(x)={1ifN(x)>0,0otherwise. (2)

The individual nuclei are defined as the connected components of the binary image Nb. This leads to an under segmentation when contours between touching nuclei are not well estimated. Consequently, we propose to apply the watershed algorithm [5] on the nuclei component N and to take advantage of the nuclei contour score Sc obtained from the U-Net or Inception-V3 output to refine the segmentation. Let us define W = {W(x)}i∈Ω the result of the watershed algorithm applied to the inverse intensity of the nuclei component N prealably convolved with a Gaussian filter of kernel size 2.5 for U-Net and 3.5 for Inception-V3. From W, a set of K new nuclei separations {S(k)}k=1,,K is obtained. A score Sw = {Sw(k)}k=1,…,K is defined for each new separation S(k) as follows:

Sw(k)=xS(k)Sc(x)-Sn(x). (3)

All nuclei separations S(k) such that Sw(k) > 0 are used to separate the nuclei obtained as the connected components of the binary image Nb.

Mask R-CNN

Version 2.1 of Mask R-CNN [31] was used in this study. The backbone network was defined as the Resnet-101 deep convolutional neural network [49]. We used the code in [7] to define the only class in this study, i.e. the nuclei. A data augmentation to increase the training dataset by a factor of 100 was processed before normalization with the imgaug python library [48] and included resizing, cropping, flipping, rotation, shearing, blurring, sharpness and brightness modifications, noise addition, intensity inversion and contrast modifications. Transfer learning with fine-tuning from a network trained on the coco dataset [33] was also applied. In the first epoch, only the region proposal network, the classifier and mask heads were trained. The whole network was then trained for the next three epochs.

Stardist

Version 0.7.1 of Stardist [8] was used in this study. The backbone network was defined as a U-Net [27]. 512 x 512 input images were normalized with a 1–99 quantile. The data augmentation proposed by the authors was used to increase the training dataset. The whole network was trained for 400 epochs.

Cellpose

Version 0.6.5 of Cellpose [9] was used in this study. The backbone network was defined as a U-Net [27]. 512 x 512 input images were normalized with a 1–99 quantile. A data augmentation to increase the training dataset was processed before normalization with the imgaug python library [48] and included flipping, rotation, shearing, blurring, sharpness and brightness modifications, noise addition, intensity inversion and contrast modifications. The whole network was trained for 400 epochs.

A summary of all the different steps used for each deep learning approach is available in S2 Table. The time computation for training and processing each individual method is shown in S1 and S4 Tables.

Evaluation

As proposed by Caicedo et al. [32], we used the F1 score with respect to the Intersection over Union (IoU) to compare the three deep learning approaches. We evaluated three different images for both confocal and widefield modalities annotated by a pathologist (a high-skilled researcher). Let OGT = {OGT(e)}e=1,…,n be the set of n ground truth nuclei, as defined by the pathologist, and OE = {OE(e)}e=1,…,m be the set of m estimated nuclei with a deep learning approach. The IoU defined between the truth nucleus OGT(e1) and the estimated nucleus OE(e2) was defined as:

IoU(OGT(e1),OE(e2))=OGT(e1)OE(e2)OGT(e1)OE(e2). (4)

An IoU(OGT(e1), OE(e2)) equal to 0 implies that OGT(e1) and OE(e2) do not share any pixel while an IoU(OGT(e1), OE(e2)) equal to 1 means that OGT(e1) and OE(e2) are identical. An IoU(OGT(e1), OE(e2)) equal to 0.5 ensures that IoU(OGT(e1), OE(e3)) < 0.5, ∀e3e2 as a nucleus cannot share half of its area with more than one nucleus. Consequently, the F1 score for a given IoU threshold t > 0.5 can be defined as:

F1(t)=2×TP(t)2×TP(t)+FN(t)+FP(t), (5)

where

TP(t)=e1{1,,n},e2{1,,m}1(IoU(OGT(e1),OE(e2))>t), (6)
FN(t)=e1{1,,n}1(IoU(OGT(e1),OE(e2))<t),e2{1,,m}, (7)
FP(t)=e2{1,,m}1(IoU(OGT(e1),OE(e2))<t),e1{1,,n}, (8)

and

1(C)={1ifCistrue,0otherwise. (9)

With a threshold t = 0.5, this metric gives the accuracy of a method to identify the correct number of nuclei, while with thresholds in the range 0.5 − 1, it evaluates the localization accuracy of the identified nuclear contours.

Marker identification

Training dataset

As Mask R-CNN and U-Net are not suited for the task, the Inception-V3 method was used for marker identification. The training dataset consisted of eight confocal images for E2F3A, five confocal images for E2F8, four confocal images for EdU, three confocal images for pH3, twenty-one widefield images for E2F3A and fifteen widefield images for E2F8, E2F4, EdU and pH3, all annotated by a moderately-skilled researcher. Ninety percent of image patches were used for training and ten percent for validation. Two training strategies were designed to define the input image patches: The first only considered image patches centered at each nucleus center, while the second strategy included image patches centered at each pixel in the segmented nuclei, drastically increasing the amount of training data.

Inception-V3

Two classes (positive and negative) were defined for the E2Fs, while three classes (diffuse, punctate and negative) were defined for EdU and pH3. As for nuclei segmentation, the imaging field was set to 65 x 65 pixel image patches to ensure inclusion of at least one entire nucleus. The training image patches were normalized by subtracting at each pixel the average intensity in a 65 x 65 neighborhood. Intensities were not divided by the median intensity over the image as they were for nuclei segmentation because the median can be equal to zero for sparse markers such as pH3. To obtain a balanced training dataset, the same number of input image patches was defined for each class. A stochastic gradient descent was used to estimate the parameters of the deep convolutional neural network by minimizing a categorical cross entropy for 10 epochs. To increase the size of the nuclei-based training dataset 100-fold, a data augmentation that included flipping, rotation, shearing, blurring, sharpness and brightness modifications, noise addition, inversion and contrast modifications was performed after normalization with the imgaug python library [48]. A transfer learning with fine-tuning from the Inception-V3 network trained for nuclei segmentation on confocal and widefield images was performed on the pixel-based training dataset. In the first epoch, only the last layer of the network was trained. The whole network was then trained for the next five epochs.

Thresholding

Marker identification was also evaluated when thresholding marker intensity in the nuclei with the ImageJ plugin Annotater. For E2Fs, nuclei with more than thirty-five percent of their area above a manually defined threshold were considered positive. For EdU, nuclei with more that seventy percent of their area above a manually defined threshold were considered diffuse, while nuclei with more than fifteen percent but less than seventy percent of their area above the same threshold were considered punctate. For pH3, nuclei with more that ninety percent of their area above a manually defined threshold were considered diffuse, while nuclei with more than ten percent but less than ninety percent of their area above the same threshold were considered punctate.

Evaluation

Evaluation was performed using three confocal and three widefield images for E2F3A and E2F8, using three widefield images for E2F4, and using two confocal and three widefield images for EdU and pH3. All the images used for evaluation were annotated by a moderately-skilled researcher. The nuclei locations were known and only their class (positive/negative or diffuse/punctate/negative) was evaluated. Let CGT = {CGT(e)}e=1,…,n be the ground truth class for the n nuclei and let CE = {CE(e)}e=1,…,n be the estimated class for each nucleus. Class estimation accuracy is then defined as:

accuracy=TPn, (10)

where

TP=e=1n1(CGT(e)=CE(e)). (11)

ImageJ plugin Annotater

We developed the plugin entitled Annotater to draw nuclear contours manually to define training datasets and to interactively correct deep learning-based nuclei segmentations. This plugin also provides for manual characterization of markers associated with nuclei and a thresholding tool for specifying an intensity threshold and the proportion of pixels in a nucleus above that threshold that defines a positive nucleus. Finally, Annotater can generate images in which segmented nuclei and marker characterization are overlaid on the input image and can extract nuclear features as defined in the Set Measurements window in ImageJ. The java code of the plugin is available at https://github.com/tpecot/Annotater/tree/master/src/main/java/edu/musc/tsl and the plugin is available at https://github.com/tpecot/Annotater. Video tutorials showing how to use the plugin are available at https://github.com/tpecot/Annotater.

E2Fs accumulation over the cell cycle from 2D still images

Data

As fluorescence intensity is proportional to protein concentration [36, 37], E2Fs nuclear accumulation over the cell cycle can be estimated from the intensity levels observed in the nuclei. To minimize the effects of noise while still being able to observe fluorescence fluctuations, we decided to define four levels of intensity, which is widely used in both diagnostic pathology and biomedical research [38, 39]. In the previous section about marker identification, positive and negative cells were manually selected to define the training and evaluation datasets. By definition, negative cells do not express proteins and were consequently assigned an intensity equal to 0. Then, for each individual image, the average fluorescence intensity for each positive cell was processed and the range of average intensities from the lowest to the highest was binned into three to define levels 1 to 3. Defining these levels from the range of average intensity observed in positive nuclei amounts to normalizing intensity over the images. S7a and S7b Fig shows the E2F8 channel for a confocal and a widefield image where the positive E2F8 nuclei are overlaid as orange circles while negative E2F8 nuclei are overlaid as red circles. The corresponding histograms for the average E2F8 intensity observed in nuclei are shown in S7c and S7d Fig. The colors in these histograms correspond to the E2F8 levels used to estimate the concentration evolution over the cell cycle. The intensity minimum and maximum for each level are different for the two images as they depend on the lowest and highest nuclear average intensities, which are different from one image to the other. However, for a given image, the level width is the same for levels 1 to 3. Additionally, some E2F8 negative nuclei show an average nuclear intensity similar to the one observed in nuclei with an intensity level of 1 (bars that are blue and yellow). This happens for nuclei with non-specific E2F8 intensity. On the other hand, the state of EdU and pH3 markers for each nucleus was already known from the marker identification as being diffuse, punctate or negative.

More formally, let us define IeCh1{0,1,2,3} the average intensity in Ch1 for nucleus e ∈ {1, …, n} in an image with n nuclei, where Ch1 is a channel corresponding to E2F3A, E2F8 or E2F4. Similarly, IeCh2{0,1,2,3} is the average intensity in Ch2 for nucleus e if Ch2 is E2F4 or E2F8 and IeCh2{0,diffuse,punctate} is the state of nucleus e if Ch2 is EdU or pH3. The estimation of fluorescence accumulation over the cell cycle is performed in 3 steps: i) initialization, ii) registration with respect to cell cycle, iii) global optimization.

Initialization

Let us define h(ICh1 = u1, ICh2 = u2) the number of nuclei for which the intensity in Ch1 is equal to u1 and the intensity or state in Ch2 is equal to u2:

h(ICh1=u1,ICh2=u2)=e=1n1(IeCh1=u1)1(IeCh2=u2). (12)

We can now define H={H(ICh1=u1,ICh2=u2)}u1={0,1,2,3},u2={0,1,2,3}oru2={0,diffuse,punctate} the 2D intensity histogram as follows:

H(ICh1=u1,ICh2=u2)=h(ICh1=u1,ICh2=u2)n. (13)

An example of a 2D histogram of E2F3A intensity and pH3 patterns is shown in S8a Fig. According to our first and second assumptions, i.e. time depends on the proportion of cells in the images and temporal evolution of protein accumulation is similar in all observed cells, 2D intensity histograms provide the proportion of nuclei for combination of E2Fs intensities or E2Fs intensity and EdU/pH3 patterns. For example, in the image considered to build the histogram shown in S8a Fig, thirty-four percent of nuclei have an intensity of 2 for E2F3A and are negative for pH3. Combining these proportions allows to define the evolution of the intensity/state of the channels over the cell cycle. We propose to initialize the evolution of fluorescence over the cell cycle through a graph optimization procedure. Let us define a graph G(E,V) with |V| vertices corresponding to the different non-zero proportions of markers from a 2D histogram and |E| edges between neighbor vertices in the same histogram. For instance, in S8a Fig, there are nine vertices {VE2F3A=0,pH3=negative,VE2F3A=1,pH3=negative,,VE2F3A=0,pH3=diffuse} represented as circles and twenty-two edges {E(VE2F3A=0,pH3=negativeVE2F3A=1,pH3=negative),,E(VE2F3A=0,pH3=diffuseVE2F3A=0,pH3=punctate)} represented as unidirectional arrows. A cost P(E(Vv1,Vv2)) is then assigned to each edge. From [14], we know that EdU is diffuse during the first half of S phase and then punctate during the second half of S phase. It implies that:

P(E(VG=u,EdU=negative,VG=u,EdU=diffuse))=0,P(E(VG=u,EdU=diffuse,VG=u,EdU=punctate))=0,P(E(VG=u,EdU=punctate,VG=u,EdU=negative))=0, (14)

while

P(E(VG=u,EdU=diffuse,VG=u,EdU=negative))=,P(E(VG=u,EdU=punctate,VG=u,EdU=diffuse))=,P(E(VG=u,EdU=negative,VG=u,EdU=punctate))=, (15)

where G ∈ {E2F3A, E2F8, E2F4} and u ∈ {0, 1, 2, 3}. From [14], we also know that pH3 is first punctate during the second half of S phase and G2 and then diffuse during mitosis, so

P(E(VG=u,pH3=negative,VG=u,pH3=punctate))=0,P(E(VG=u,pH3=punctate,VG=u,pH3=diffuse))=0,P(E(VG=u,pH3=diffuse,VG=u,pH3=negative))=0,P(E(VG=u,pH3=punctate,VG=u,pH3=negative))=,P(E(VG=u,pH3=diffuse,VG=u,pH3=punctate))=,P(E(VG=u,pH3=negative,VG=u,pH3=diffuse))=, (16)

where G ∈ {E2F3A, E2F8, E2F4} and u ∈ {0, 1, 2, 3}. Finally, the evolution of E2Fs intensity needs to verify our third assumption, i.e. concentrations of E2Fs are concave downward parabolas, which explains why edges exist only between neighboring vertices in the 2D histograms:

P(E(VG=u1,M=s,VG=u2,M=s))={0if|u1-u2|=1,otherwise, (17)

where G ∈ {E2F3A, E2F8, E2F4}, u1 ∈ {0, 1, 2, 3}, u2 ∈ {0, 1, 2, 3}, M ∈ {EdU, pH3} and s ∈ {negative, punctate, diffuse}. The evolution of the fluorescence is initialized as the sequence of edges that goes through all vertices with a minimum cost. For instance, the only sequence of edges with a cost equal to 0 in the 2D histogram shown in S8a Fig is {VE2F3A=0,pH3=negative,VE2F3A=1,pH3=negative,VE2F3A=2,pH3=negative,VE2F3A=3,pH3=negative,VE2F3A=3,pH3=punctate,VE2F3A=2,pH3=punctate,VE2F3A=1,pH3=punctate,VE2F3A=0,pH3=punctate,VE2F3A=0,pH3=diffuse}. This sequence corresponds to the evolution of E2F3A and pH3 shown in S8b Fig.

We know that E2F4 shows two waves of expression over the cell cycle [14], but a sequence of edges over a 2D histogram cannot reflect two downward parabolas. Consequently, splitting vertex(ices) is necessary. A cost of 1 is associated to each vertex splitting and added to the total cost. The split vertices give rise to multiple vertices containing equal proportion. For combinations of E2Fs with EdU as well as combinations of two different E2Fs, vertex splitting can also be required and may lead to E2Fs evolutions that do not look exactly like downward parabolas. This phenomenon arises from slight errors in nuclear intensity measurements, most likely due to image noise. A 2D histogram for the combination of E2F3A and E2F4 is shown in S9a Fig, the sequence of edges over this histogram for which vertices were split is shown in S9b Fig and the corresponding evolution of E2F3A and E2F4 is shown in S9c Fig.

By applying this cost minimization to all images, initializations for the eight different combinations of markers and averaged for each mouse are obtained. These initializations are shown in S10 Fig. From these initializations, eight couples of marker sequences are defined:

{initXiEdU-E2F3A(c),initXiE2F3A-EdU(c)}c=1,,B,{initXipH3-E2F3A(c),initXiE2F3A-pH3(c)}c=1,,B,{initXiEdU-E2F8(c),initXiE2F8-EdU(c)}c=1,,B,{initXipH3-E2F8(c),initXiE2F8-pH3(c)}c=1,,B,{initXiEdU-E2F4(c),initXiE2F4-EdU(c)}c=1,,B,{initXipH3-E2F4(c),initXiE2F4-pH3(c)}c=1,,B,{initXiE2F3A-E2F8(c),initXiE2F8-E2F3A(c)}c=1,,B,{initXiE2F3A-E2F4(c),initXiE2F4-E2F3A(c)}c=1,,B, (18)

where initXiM1-M2(c) provides the level of fluorescence or pattern for marker M1 ∈ {E2F3A, E2F8, E2F4, EdU, pH3} when combined with marker M2 ∈ {E2F3A, E2F8, E2F4, EdU, pH3} for mouse i at time c, M1 ≠ M2, and B is the number of time bins used to quantize the cell cycle: 100 bins in practice. The goal is to estimate the sequence for each of these couples of combinations.

Registration

As EdU and pH3 expression patterns with respect to the cell cycle are known, E2Fs are first registered with respect to these two markers. We know that pH3 is diffuse during mitosis. Consequently, this pattern is considered as the reference and each initialization of any combination with pH3 ends with a diffuse pH3 pattern corresponding to the end of the cell cycle. This implies that the pH3 sequences are known and defined as XiG-pH3 for each mouse i. A circular permutation is operated next to estimate the XiEdU-G where G ∈ {E2F3A, E2F8, E2F4} as follows:

min{0-XiEdU-G,0-XiG-EdU}j=1NG0-XiG-EdU-XjG-pH322,subjectto0-XiG-EdU(c)=initXiG-EdU(c-q),q{1,,B},i{1,,NG},G{E2F3A,E2F8,E2F4}, (19)

where NG = 3 for E2F3A and E2F8, NG = 2 for E2F4, and 0-XiG-EdU(c-q)=initXiG-EdU(B-(c-q)) if (cq) is negative. We solve this optimization problem by computing j=1NG0-XiG-EdU-XjG-pH322 for all possible shifts c = {1, …, B}. The solution corresponds to the shift c that gives the minimum sum. To better estimate EdU and initialize E2Fs, we now use the E2Fs expression:

min{0XiEdU-G1,0XiG1-EdU}j=1NG10XiG1-EdU-XjG1-pH322+j=1NG20XiEdU-G1-0XjEdU-G222+j=1NG30XiEdU-G1-0XjEdU-G322,subjectto0XiEdU-G1(c)=0-XiEdU-G1(c-q),q{1,,B},i{1,,NG1}

where {G1, G2, G3} are successively {E2F3A, E2F8, E2F4}, {E2F8, E2F3A, E2F4} and {E2F4, E2F8, E2F3A}, NGi = 3 if Gi is E2F3A or E2F8 and NGi = 2 if Gi is E2F4. Again, we solve this optimization problem by computing the distances (20) for all possible shifts c = {1, …, B} and obtain the solution with the minimum value. Now, the temporal evolutions of EdU XiEdU-G and pH3 XipH3-G are known for each mouse i and each genotype G ∈ {E2F3A, E2F8, E2F4}. This implies that EdU and pH3 evolutions will not change but can be permuted for a given state of EdU or pH3. The E2Fs concentrations and patterns for EdU and pH3 after registration are shown in S11 Fig.

Assignment problem

Now that EdU and pH3 are known, we define the estimation of E2Fs as an assignment problem where the goal is to find the permutations for each XiG-M,G{E2F3A,E2F8,E2F4} and M ∈ {EdU, pH3} such that E2Fs accumulation over time are the most similar across data while EdU and pH3 are already known. We model this estimation as an assignment problem and define the following square cost matrices:

Cost((z)XiG1-M1(c1,c2))=j=1NG1(z-1)XiG1-M1(c1)-(z)XjG1-M2(c2)22+j=1NG2(z-1)XiG1-M1(c1)-(z)XjG1-G2(c2)22+j=1NG3(z-1)XiG1-M1(c1)-(z)XjG1-G3(c2)22+(z)XiM1-G1(c1)-XiM1-G1(c2)+(c1-c2)2z,i{1,,NG1}, (20)

where the norm is used to ensure that EdU and pH3 states stay the same over the cell cycle, {G1, G2, G3} are successively {E2F3A, E2F8, E2F4}, {E2F8, E2F3A, E2F4} and {E2F4, E2F8, E2F3A}, NGi = 3 if Gi is E2F3A or E2F8 and NGi = 2 if Gi is E2F4, {M1, M2} are successively {EdU, pH3} and {pH3, EdU}, z corresponds to the iteration of the algorithm and:

Cost((z)XiE2F3A-G1(c1,c2))=j=13{(z-1)XiE2F3A-G1(c1)-(z-1)XjE2F3A-EdU(c2)22+(z-1)XiE2F3A-G1(c1)-(z-1)XjE2F3A-pH3(c2)22}+j=1NG1{(z-1)XiG1-E2F3A(c1)-(z-1)XjG1-EdU(c2)22+(z-1)XiG1-E2F3A(c1)-(z-1)XjG1-pH3(c2)22}+j=1NG2(z-1)XiE2F3A-G1(c1)-(z-1)XjE2F3A-G2(c2)22+(c1-c2)2z,i{1,,NG1}, (21)

where {G1, G2} are successively {E2F8, E2F4} and {E2F4, E2F8}, NGi = 3 if Gi is E2F8 and NGi = 2 if Gi is E2F4. Each one of these five minimizations is processed one after another for iterations z = {1, …, B} by using the Hungarian algorithm [24, 25]. In Eq (20), the three first lines reflect the goal to have a given E2F accumulation over the cell cycle as similar as possible when combined with EdU, pH3 or another E2F; the fourth line corresponds to the constraint that EdU and pH3 are known and their state over the cell cycle cannot change, even though there can be permutations for a given state; and the last line defines a local constraint that forces permutations to be restricted in time for the first iterations and allows them to grow when iterations get larger. The latter constraint is used because the data is well initialized so the final estimation should not permute points directly all over the cell cycle. Eq (21) takes into account the differences between E2F3A accumulation over the cell cycle when combined with any marker or any other E2F, with the same constraint as in Eq (20). Finally, a spline fitting is applied to each of the E2Fs across all experiments to get a final accumulation over the cell cycle, which is shown in Fig 3. The code with the parameters used to estimate E2Fs accumulation over the cell cycle after initialization is available at https://github.com/tpecot/EstimationOfProteinConcentrationOverTime.

Simulated data

We generated simulations to evaluate the accuracy of the temporal estimation of protein concentration. For each simulation, the following concentrations are generated:

yiE2F3Aj{0,,nbI},i{0,,nbT},j{0,,nbS},yiE2F8j{0,,nbI},i{0,,nbT},j{0,,nbS},yiE2F4j{0,,nbI},i{0,,nbT},j{0,,nbS2},yiEdUj{0,diffuse,punctate},i{0,,nbT},j{0,,nbS},yipH3j{0,punctate,diffuse},i{0,,nbT},j{0,,nbS}, (22)

where nbI is the number of bins for intensity, nbT is the number of bins for time and nbS is the number of samples for E2F3A, E2F8, EdU and pH3 while nbS2 is the number of samples for E2F4, corresponding to mice in our study. YE2F3A and YE2F8 are simulated as concentrations starting at intensity 0 to nbI, and then back to 0 with a duration for each intensity bin randomly generated between 1 and 23*nbI*nbT to ensure that the concentration is shorter that the simulation. The starting point for these concentrations is also randomly generated. YE2F4 is simulated as two waves of concentrations starting at intensity 0 to nbI, and then back to 0 with a duration for each intensity bin randomly generated between 1 and 13*nbI*nbT to ensure that both waves can fit in the simulation duration. The starting point for both waves are randomly generated, with the constraint that the second wave starts after the first one ends. YEdU and YpH3 are simulated as successive diffuse and punctate patterns for EdU and successive punctate and diffuse patterns for pH3, with a duration for each pattern randomly generated between 1 and 23*nbI*nbT. The starting point for these markers is also randomly generated, with the constraint that pH3 starts after EdU. To evaluate the robustness of our approach, noise was randomly added to affect a range from 0% to 50% of corrupted time bins. In addition, five different simulations were randomly generated for each set of parameters. The performance was evaluated by measuring the mean squared error (MSE) between the generated and estimated concentrations of E2Fs, EdU and pH3.

Three different scenarios were designed to evaluate the performance of our approach. In a first set of experiments, the number of samples was evaluated and defined as {n1 = 1; n2 = 1}, {n1 = 2; n2 = 2}, {n1 = 3; n2 = 2}, {n1 = 3; n2 = 3} and {n1 = 5; n2 = 5}, with nbI = 4 and nbT = 100. The results are shown in S12 Fig. Then, the influence of the number of time bins was estimated by considering nbT = {20, 50, 75, 100, 125, 150, 200} with {n1 = 3; n2 = 2} and nbI = 4. The results are shown in S13 Fig. Finally, the impact of the number of intensity bins was evaluated by defining nbI = {2, 3, 4, 5, 10, 20} with {n1 = 3; n2 = 2} and nbT = 100. The results are shown in S14 Fig.

Supporting information

S1 Fig. Example images.

One example for each combination of markers and modality for fluorescence images used in the study. The combination of markers, number of images of each type, modality, and number of mice used in the study are shown on top of the images. Note the small size of the training set. Scale bar = 20μm.

(TIF)

S2 Fig. Evaluation of U-Net for nuclei segmentation.

F1 score for multiple IoU thresholds obtained with the U-Net approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation or post-processing a-b, with data augmentation c-d, with data augmentation and corrected watershed postprocessing e-f. The lines correspond to the average F1 score while the areas represent the standard error.

(TIF)

S3 Fig. Evaluation of Inception-V3 for nuclear segmentation.

F1 score for multiple IoU thresholds obtained with the Inception-V3 approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation or post-processing a-b, with data augmentation c-d, with data augmentation and corrected watershed postprocessing e-f. The lines correspond to the average F1 score while the areas represent the standard error.

(TIF)

S4 Fig. Evaluation of Mask R-CNN for nuclear segmentation.

F1 score for multiple IoU thresholds obtained with the Mask R-CNN approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation but with transfer learning a-b, without transfer learning but with data augmentation c-d, with data augmentation and transfer learning e-f. The lines correspond to the average F1 score while the areas represent the standard error.

(TIF)

S5 Fig. Evaluation of Stardist for nuclear segmentation.

F1 score for multiple IoU thresholds obtained with the Stardist approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation a-b and with data augmentation c-d. The lines correspond to the average F1 score while the areas represent the standard error.

(TIF)

S6 Fig. Evaluation of Cellpose for nuclear segmentation.

F1 score for multiple IoU thresholds obtained with the Cellpose approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation a-b and with data augmentation c-d. The lines correspond to the average F1 score while the areas represent the standard error.

(TIF)

S7 Fig. Examples of intensity binning for E2F8.

a-b E2F8 channel of a confocal a and a widefield b image. E2F8 positive nuclei are overlaid as orange circles while E2F8 negative nuclei are overlaid as red circles. Scale bar = 20μm. c-d Histograms of the E2F8 intensity observed in the nuclei shown in images a (c) and b (d). The 3 levels of intensity used for concentration estimation are displayed with different colors.

(TIF)

S8 Fig. Initialization of E2F3A and pH3 over time.

a 2D histogram of E2F3A intensity and pH3 patterns. b Initialization of E2F3A and pH3 over time from the 2D histogram shown in a.

(TIF)

S9 Fig. Initialization of E2F3A and E2F4 over time.

a 2D histogram of E2F3A and E2F4 intensity. b Modified 2D histogram of E2F3A and E2F4 intensity with vertex splitting and set of edges giving the minimum cost. c Initialization of E2F3A and E2F4 over time from the set of edges shown in b.

(TIF)

S10 Fig. Initialization of E2Fs with respect to EdU and pH3.

a-b Initialization of E2F3A with respect to a EdU (n = 3) and b pH3 (n = 3). c-d Initialization of E2F8 concentration with respect to c EdU (n = 3) and d pH3 (n = 3). e-f Initialization of E2F4 concentration with respect to e EdU (n = 2) and f pH3 (n = 2). g-h Initialization of E2F3A concentration with respect to g E2F8 (n = 3) and h E2F4 (n = 2). For all curves, the E2Fs intensity for the different mice are represented as curves with different line styles while EdU and pH3 are shown above the E2Fs curves, with solid lines corresponding to diffuse states and dashed lines corresponding to punctate states.

(TIF)

S11 Fig. Initialization of E2Fs evolution over the cell cycle.

a Temporal evolution of E2F3A over the cell cycle after registration between EdU and pH3 (n = 6). b Temporal evolution of E2F8 over the cell cycle after registration between EdU and pH3 (n = 6). c Temporal evolution of E2F4 over the cell cycle after registration between EdU and pH3 (n = 4). For all curves, the E2Fs intensity for the different mice are represented as curves with different line styles while EdU and pH3 are shown above the E2Fs curves, with solid lines corresponding to diffuse states and dashed lines corresponding to punctate states.

(TIF)

S12 Fig. Evaluation of the influence of the number of samples over the estimation of protein concentration over the cell cycle on simulated data.

Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of samples when corrupting up to 50% of the simulated data with noise. n1 corresponds to the number of samples for E2F3A, E2F8, EdU and pH3 while n2 is the number of samples for E2F4. The lines correspond to the average mean squared error while the areas represent the standard error.

(TIF)

S13 Fig. Evaluation of the influence of the number of time bins over the estimation of protein concentration over the cell cycle on simulated data.

Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of time bins when corrupting up to 50% of the simulated data with noise. The lines correspond to the average mean squared error while the areas represent the standard error.

(TIF)

S14 Fig. Evaluation of the influence of the number of intensity bins over the estimation of protein concentration over the cell cycle on simulated data.

Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of intensity bins when corrupting up to 50% of the simulated data with noise. The lines correspond to the average mean squared error while the areas represent the standard error.

(TIF)

S1 Table. Time computation for training nuclei segmentation models.

Computation time needed to train the five different deep learning approaches on the four training datasets with a GeForce RTX 2080 with Max-Q design. CF stands for confocal images, WF half stands for half the widefield images, WF stands for widefield images and CFWF stands for confocal and widefield images.

(TIF)

S2 Table. Training strategies for nuclei segmentation.

Pre-processing, normalization, optimization method, data augmentation, transfer learning and post-processing used to train the five deep learning approaches.

(TIF)

S3 Table. Training data summary.

Number of images and nuclei in the four different training and validation datasets. CF stands for confocal images, WF half stands for half the widefield images, WF stands for widefield images and CFWF stands for confocal and widefield images.

(TIF)

S4 Table. Time computation for processing nuclei segmentation.

Computation time needed to process the five different deep learning approaches on the four training datasets with a GeForce RTX 2080 with Max-Q design.

(TIF)

S5 Table. Evaluation of deep learning approaches for nuclei segmentation in confocal images.

F1 score obtained for IoU = 0.5 and IoU = 0.75 when segmenting nuclei in confocal images with U-Net, Inception-V3, Mask R-CNN, Stardist and Cellpose. DA stands for data augmentation, PP stands for post-processing, TL stands for transfer learning and IoU stands for intersection over union.

(TIF)

S6 Table. Evaluation of deep learning approaches for nuclei segmentation in widefield images.

F1 score obtained for IoU = 0.5 and IoU = 0.75 when segmenting nuclei in widefield images with U-Net, Inception-V3, Mask R-CNN, Stardist and Cellpose. DA stands for data augmentation, PP stands for post-processing, TL stands for transfer learning and IoU stands for intersection over union.

(TIF)

S7 Table. Performance of Inception-V3 for marker identification with confocal images.

Inception-V3 accuracy and standard error with and without transfer learning for marker identification of E2F3A, E2F8, EdU and pH3 in confocal images. DA stands for data augmentation.

(TIF)

S8 Table. Performance of Inception-V3 for marker identification with widefield images.

Inception-V3 accuracy and standard error with and without transfer learning for marker identification of E2F3A, E2F8, E2F4, EdU and pH3 in widefield images. DA stands for data augmentation.

(TIF)

S9 Table. Time computation for training and processing marker identification.

Computation time needed to train and process Inception-V3 for marker identification with a GeForce RTX 2080 with Max-Q design. DA stands for data augmentation.

(TIF)

Data Availability

Data availability All the images used in this study are available at https://data.mendeley.com/datasets/5r6kf37zd4/1. The training datasets for nuclei segmentation are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/UNet/datasets/nucleiSegmentation_E2Fs for the U-Net architecture and at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/MaskRCNN/datasets/nucleiSegmentation_E2Fs for the Mask R-CNN architecture. The training datasets for nuclei segmentation and marker identification are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/InceptionV3/trainingData for the Inception-V3 architecture. The images used for the evaluation and the ground truth are available at the same locations as the training datasets. The intensity 2D histograms used to estimate the E2Fs accumulation over the cell cycle are available at https://github.com/tpecot/EstimationOfProteinConcentrationOverTime/tree/master/data. Software availability The codes used to train and process deep learning approaches for nuclei segmentations and marker identification are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification. Archived code at time of publication: https://doi.org/10.5281/zenodo.4619243 [50] License: GPL3 The Octave code used to estimate the E2Fs accumulation over the cell cycle is available at https://github.com/tpecot/EstimationOfProteinConcentrationOverTime. Archived code at time of publication: https://doi.org/10.5281/zenodo.4639800 [51] License: GPL3 The Java code of the Annotater plugin and the plugin are available at https://github.com/tpecot/Annotater. Video tutorials to show how to use the Annotater are available at the same location. Archived code at time of publication: https://doi.org/10.5281/zenodo.4639802 [52] License: GPL3.

Funding Statement

This work was funded by a Chan Zuckerberg Initiative (https://chanzuckerberg.com/) DAF grant to T.P. (2019-198009), an NCI grant (R50 CA211529) to C.T., Advancing a Healthier Wisconsin Endowment to G.L. and a Dr. Glenn R. and Nancy A. Linnerson Endowed Fund to G.L.. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems; 2012. p. 1097–1105.
  • 2.Cireşan D, Meier U, Schmidhuber J. Multi-column deep neural networks for image classification. In: 2012 IEEE conference on computer vision and pattern recognition. IEEE; 2012. p. 3642–3649.
  • 3. LeCun Y, Bengio Y, Hinton G. Deep learning. nature. 2015;521(7553):436–444. doi: 10.1038/nature14539 [DOI] [PubMed] [Google Scholar]
  • 4. Schmidhuber J. Deep learning in neural networks: An overview. Neural networks. 2015;61:85–117. doi: 10.1016/j.neunet.2014.09.003 [DOI] [PubMed] [Google Scholar]
  • 5. Vincent L, Soille P. Watersheds in digital spaces: an efficient algorithm based on immersion simulations. IEEE Transactions on Pattern Analysis & Machine Intelligence. 1991;(6):583–598. doi: 10.1109/34.87344 [DOI] [Google Scholar]
  • 6. Moen E, Bannon D, Kudo T, Graf W, Covert M, Van Valen D. Deep learning for cellular image analysis. Nature methods. 2019; p. 1–14. doi: 10.1038/s41592-019-0403-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Hollandi R, Szkalisity A, Toth T, Tasnadi E, Molnar C, Mathe B, et al. nucleAIzer: a parameter-free deep learning framework for nucleus segmentation using image style transfer. Cell Systems. 2020;10(5):453–458. doi: 10.1016/j.cels.2020.04.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Schmidt U, Weigert M, Broaddus C, Myers G. Cell detection with star-convex polygons. In: International Conference on Medical Image Computing and Computer-Assisted Intervention; 2018. p. 265–273.
  • 9. Stringer C, Wang T, Michaelos M, Pachitariu M. Cellpose: a generalist algorithm for cellular segmentation. Nature Methods. 2021;18(1):100–106. doi: 10.1038/s41592-020-01018-x [DOI] [PubMed] [Google Scholar]
  • 10. Chen HZ, Tsai SY, Leone G. Emerging roles of E2Fs in cancer: an exit from cell cycle control. Nature Reviews Cancer. 2009;9(11):785–797. doi: 10.1038/nrc2696 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Iaquinta PJ, Lees JA. Life and death decisions by the E2F transcription factors. Current opinion in cell biology. 2007;19(6):649–657. doi: 10.1016/j.ceb.2007.10.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Lammens T, Li J, Leone G, De Veylder L. Atypical E2Fs: new players in the E2F transcription factor family. Trends in cell biology. 2009;19(3):111–118. doi: 10.1016/j.tcb.2009.01.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Logan N, Graham A, Zhao X, Fisher R, Maiti B, Leone G, et al. E2F-8: an E2F family member with a similar organization of DNA-binding domains to E2F-7. Oncogene. 2005;24(31):5000–5004. doi: 10.1038/sj.onc.1208703 [DOI] [PubMed] [Google Scholar]
  • 14. Cuitiño MC, Pécot T, Sun D, Kladney R, Okano-Uchida T, Shinde N, et al. Two distinct E2F transcriptional modules drive cell cycles and differentiation. Cell reports. 2019;27(12):3547–3560. doi: 10.1016/j.celrep.2019.05.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Humbert PO, Rogers C, Ganiatsas S, Landsberg RL, Trimarchi JM, Dandapani S, et al. E2F4 is essential for normal erythrocyte maturation and neonatal viability. Molecular cell. 2000;6(2):281–291. doi: 10.1016/S1097-2765(00)00029-0 [DOI] [PubMed] [Google Scholar]
  • 16. Li J, Ran C, Li E, Gordon F, Comstock G, Siddiqui H, et al. Synergistic function of E2F7 and E2F8 is essential for cell survival and embryonic development. Developmental cell. 2008;14(1):62–75. doi: 10.1016/j.devcel.2007.10.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Rempel RE, Saenz-Robles MT, Storms R, Morham S, Ishida S, Engel A, et al. Loss of E2F4 activity leads to abnormal development of multiple cellular lineages. Molecular cell. 2000;6(2):293–306. doi: 10.1016/S1097-2765(00)00030-7 [DOI] [PubMed] [Google Scholar]
  • 18. Tsai SY, Opavsky R, Sharma N, Wu L, Naidu S, Nolan E, et al. Mouse development with a single E2F activator. Nature. 2008;454(7208):1137–1141. doi: 10.1038/nature07066 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Van Valen DA, Kudo T, Lane KM, Macklin DN, Quach NT, DeFelice MM, et al. Deep learning automates the quantitative analysis of individual cells in live-cell imaging experiments. PLoS computational biology. 2016;12(11). doi: 10.1371/journal.pcbi.1005177 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Blom S, Paavolainen L, Bychkov D, Turkki R, Mäki-Teeri P, Hemmes A, et al. Systems pathology by multiplexed immunohistochemistry and whole-slide digital image analysis. Scientific reports. 2017;7(1):1–13. doi: 10.1038/s41598-017-15798-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Keren L, Bosse M, Marquez D, Angoshtari R, Jain S, Varma S, et al. A structured tumor-immune microenvironment in triple negative breast cancer revealed by multiplexed ion beam imaging. Cell. 2018;174(6):1373–1387. doi: 10.1016/j.cell.2018.08.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Schneider CA, Rasband WS, Eliceiri KW. NIH Image to ImageJ: 25 years of image analysis. Nature methods. 2012;9(7):671–675. doi: 10.1038/nmeth.2089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Schindelin J, Arganda-Carreras I, Frise E, Kaynig V, Longair M, Pietzsch T, et al. Fiji: an open-source platform for biological-image analysis. Nature methods. 2012;9(7):676–682. doi: 10.1038/nmeth.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Kuhn HW. The Hungarian method for the assignment problem. Naval research logistics quarterly. 1955;2(1-2):83–97. doi: 10.1002/nav.3800020109 [DOI] [Google Scholar]
  • 25. Munkres J. Algorithms for the assignment and transportation problems. Journal of the society for industrial and applied mathematics. 1957;5(1):32–38. doi: 10.1137/0105003 [DOI] [Google Scholar]
  • 26.Cireşan DC, Giusti A, Gambardella LM, Schmidhuber J. Mitosis detection in breast cancer histology images with deep neural networks. In: International conference on medical image computing and computer-assisted intervention. Springer; 2013. p. 411–418. [DOI] [PubMed]
  • 27.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer; 2015. p. 234–241.
  • 28. Caicedo JC, Goodman A, Karhohs KW, Cimini BA, Ackerman J, Haghighi M, et al. Nucleus segmentation across imaging experiments: the 2018 Data Science Bowl. Nature methods. 2019;16(12):1247–1253. doi: 10.1038/s41592-019-0612-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Falk T, Mai D, Bensch R, Çiçek Ö, Abdulkadir A, Marrakchi Y, et al. U-Net: deep learning for cell counting, detection, and morphometry. Nature methods. 2019;16(1):67–70. doi: 10.1038/s41592-018-0261-2 [DOI] [PubMed] [Google Scholar]
  • 30.Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 2818–2826.
  • 31.He K, Gkioxari G, Dollár P, Girshick R. Mask r-cnn. In: Proceedings of the IEEE international conference on computer vision; 2017. p. 2961–2969.
  • 32. Caicedo JC, Roth J, Goodman A, Becker T, Karhohs KW, Broisin M, et al. Evaluation of deep learning strategies for nucleus segmentation in fluorescence images. Cytometry Part A. 2019;95(9):952–965. doi: 10.1002/cyto.a.23863 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Lin TY, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al. Microsoft coco: Common objects in context. In: European conference on computer vision. Springer; 2014. p. 740–755.
  • 34. Gut G, Tadmor MD, Pe’er D, Pelkmans L, Liberali P. Trajectories of cell-cycle progression from fixed cell populations. Nature methods. 2015;12(10):951–954. doi: 10.1038/nmeth.3545 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Eulenberg P, Köhler N, Blasi T, Filby A, Carpenter AE, Rees P, et al. Reconstructing cell cycle and disease progression using deep learning. Nature communications. 2017;8(1):1–6. doi: 10.1038/s41467-017-00623-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Toki MI, Cecchi F, Hembrough T, Syrigos KN, Rimm DL. Proof of the quantitative potential of immunofluorescence by mass spectrometry. Laboratory Investigation. 2017;97(3):329–334. doi: 10.1038/labinvest.2016.148 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Lopes N, Bergsland CH, Bjørnslett M, Pellinen T, Svindland A, Nesbakken A, et al. Digital image analysis of multiplex fluorescence IHC in colorectal cancer recognizes the prognostic value of CDX2 and its negative correlation with SOX2. Laboratory Investigation. 2020;100(1):120–134. doi: 10.1038/s41374-019-0336-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Lindeman GJ, Gaubatz S, Livingston DM, Ginsberg D. The subcellular localization of E2F-4 is cell-cycle dependent. Proceedings of the National Academy of Sciences. 1997;94(10):5095–5100. doi: 10.1073/pnas.94.10.5095 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Maiti B, Li J, De Bruin A, Gordon F, Timmers C, Opavsky R, et al. Cloning and characterization of mouse E2F8, a novel mammalian E2F family member capable of blocking cellular proliferation. Journal of Biological Chemistry. 2005;280(18):18211–18220. doi: 10.1074/jbc.M501410200 [DOI] [PubMed] [Google Scholar]
  • 40. Campton DE, Ramirez AB, Nordberg JJ, Drovetto N, Clein AC, Varshavskaya P, et al. High-recovery visual identification and single-cell retrieval of circulating tumor cells for genomic analysis using a dual-technology platform integrated with automated immunofluorescence staining. BMC cancer. 2015;15(1):360. doi: 10.1186/s12885-015-1383-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Pitarresi JR, Liu X, Sharma SM, Cuitiño MC, Kladney RD, Mace TA, et al. Stromal ETS2 regulates chemokine production and immune cell recruitment during acinar-to-ductal metaplasia. Neoplasia. 2016;18(9):541–552. doi: 10.1016/j.neo.2016.07.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Van der Walt S, Colbert SC, Varoquaux G. The NumPy array: a structure for efficient numerical computation. Computing in Science & Engineering. 2011;13(2):22–30. doi: 10.1109/MCSE.2011.37 [DOI] [Google Scholar]
  • 43.Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems; 2015. Available from: https://www.tensorflow.org/.
  • 44.Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. Advances in Neural Information Processing Systems 32. 2019; p. 8024–8035.
  • 45.Chollet F. Keras; 2015. Available from: https://github.com/fchollet/keras.
  • 46. Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods. 2020. doi: 10.1038/s41592-020-0772-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Van der Walt S, Schönberger JL, Nunez-Iglesias J, Boulogne F, Warner JD, Yager N, et al. scikit-image: image processing in Python. PeerJ. 2014;2:e453. doi: 10.7717/peerj.453 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Jung AB, Wada K, Crall J, Tanaka S, Graving J, Reinders C, et al. imgaug; 2020. Available from: https://github.com/aleju/imgaug.
  • 49.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.
  • 50.Pécot T. Nuclei segmentation and marker identification; 2021. Available from: 10.5281/zenodo.5878923. [DOI]
  • 51.Pécot T. Estimation of protein concentration over time; 2021. Available from: 10.5281/zenodo.4639800. [DOI]
  • 52.Pécot T. Annotater; 2022. Available from: 10.5281/zenodo.5878946. [DOI]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009949.r001

Decision Letter 0

Jason M Haugh, Attila Csikász-Nagy

4 Jun 2021

Dear Dr Pecot,

Thank you very much for submitting your manuscript "Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Attila Csikász-Nagy

Associate Editor

PLOS Computational Biology

Jason Haugh

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors have three main contributions.

1- Benchmarking three existing computer vision architectures for cell segmentation (U-Net, Mask-R-CNN, Inception V3)

2- Benchmarking a computer vision architectures for estimating protein concentration (Inception V3)

3- Software tools that implement their pipeline (github repo + imagej plugin).

4- An optimization based approach to estimate protein concentration over time.

In part 1 and 2, the results are presented in an extremely confusing way. For example:

- At the very least, the authors should have a table, showing most of the relevant information in a single place. This would include: number of cells (not just images!) used for training, training time, prediction, etc. Figure 2 includes 2 subplots to show 16 numbers.

- It would be nice to see training curves for the different methods. The authors speculate that certain methods benefit more from data, but it would be nice to see this evaluated.

-Computer vision is a very quickly moving field, and I don't expect people in other fields to use and benchmark all the latest methods - but - it would be nice to see more methods get benchmarked. Ross Wightman has all the major methods implemented here with a consistent API and pretrained weights: https://github.com/rwightman/pytorch-image-models

- The methods are not described in nearly enough detail, and the text is difficult to follow. We should have clear diagrams that show the training pipeline for each method, which method used pre-training, which didn't, what method used post processing, etc - some of this information is scattered throughout the text, but it's a huge pain to have to go back and forth.

- The Python code (as presented in the repo) is not really packaged in a way that someone else could use it for their own use case. I appreciate the authors making the code available - but - in its current form, it's not really a resource for anyone else. At the very least, there should be a better Readme. Ideally, the code would be packaged in a library with a unified API.

- The authors discuss quantizing E2F signal in 4 buckets, but never show any kind of data justifying this decision - at the very least a histogram of concentration should be shown.

- The last part of the paper showing how the graph assignment is solved is clever, and I appreciated the author explaining why they felt justified in making certain assumptions. I wish they slightly more precise - for example - what exactly does "2 Temporal evolution of protein accumulation is similar in all observed cells." mean?

Overall, I feel the technical issues in the paper + the lack of clarity means the paper should not be accepted in its current form.

Reviewer #2: The manuscript focuses on the temporal expression of E2F proteins during the cell cycle.

In a first part, the segmentation of nuclei is performed based on a deep-learning

approach. The performance of several networks are compared, and the Mask R-CNN is chosen as

more appropriate. In a second part, the fluorescence pattern of proteins is investigated,

using a combination of deep-learning approach and manual correction of the results.

A software was developed to help correct and validate the results obtained by automated

segmentation. Finally, the temporal evolution of protein concentration is investigated by

considering a large population of cells at various stages in a static image, and reconstructing

the time variations of fluorescence between stages using graph optimization methods.

The manuscript present a number of results together with an exhaustive explanation on the

methods. However, I found it difficult to read, making it complicated to clearly identify what

were the questions addressed within the text. For instance, it is not clear if the paper aims at

solving a biological question by using computational methods, or if it focuses on the

methodological part, in particular deep-learning segmentation, using the biological question as

an illustration. This confusion deserves the presentation of the results, that however seem to

present interest.

If the biology is the main concern, I suggest putting more emphasis on it in the introduction,

and introduce the methodological questions based on it. One possibility could be to start from

the limitations encountered in ref [11], and use them to introduce the proposed enhancements.

Another option could be to focus on the methodological questions. Actually the results put

emphasis on the performance and accuracy of the methods. However this option would imply spliting

the manuscript into two distinct papers.

Other comments

* the introduction reviews segmentation of fluorescent structures approaches by considering

deep-learning approaches. While very popular, DL-approach is not the only approach for

segmentation, and discussing with respect to other "classical" approaches could help relate to

the field.

* L50-52: well, I guess everyone would expect a deep-learning approach to perform better than

a threshold approach...

* L56: it would be more clear to explain the three assumptions directly in the introduction.

* L92 (related to first point) authors use watershed to improve their segmentation workflow.

WS could be stated in the introduction.

* L116: It is not very clear to me why the authors use two different modalities. Is this to

increase the genericiy of the segmentation approach? the strategy should be better explained.

* L118 better justify relevancy of transfer learning (and of using both confocal and wide-field).

If the structures to segment are not the same, it is rather surprising to apply the same network.

Also, it is not natural to discover the use of transfer learning at his point, this could have

been stated earlier.

* L159: sentence is not clear.

* L160: this section would benefit from a better explanation of what do the author want to

obtain from the image, including an explanation / a recall about the different patterns (punctate

or diffuse) that can be observed during cycle, and eventually the associated segmentation difficulties.

* L160: I understand that thresholding may be limited for segmentation of diffuse patterns.

But many image segmentation methods exist and could have been explored. It would be interesting to

better explain the difficulty encountered, and better justify why DL seems to be the most adequate.

* L177: it is not clear if the "Annotater" plugin is a side-product of the research, or if the

whole workflow rely on it. This could be clarified.

* L204: maybe recall why it is interesting to focus on protein concentration. the transition from

previous paragraph is abrupt.

* L 236: concentrations are not parabola... The representation of their evolution may depict a

parabola, however.

* L360-367: the note at the nd of the paragraph would be more appropriate in the result or in the

discussion section.

* L492-615: this section if quite technical, and it is rather surprising to have such an inbalance

between the method description and the exploitation of the results. I wonder if this modelling part

can not be an article of itself?

Reviewer #3: This manuscript by T. Pecot and al. propose a workflow, based on deep learning (DL) segmentation and modeling, to reconstruct time courses of protein accumulation from fixed images. This work is timely and important for two reasons. First, while DL has been shown to work well for biological images, most reported uses consider proof of principle, standard datasets, or simple cases; reporting usage in full fledged, biologically relevant, and non trivial example is important and useful. Second in vivo video microscopy is costly and difficult, when it is possible, and recovering time courses from fixed samples in a reliable and efficient way is also important and useful. Overall the paper is well written, the study is well conducted, and the methods used, evaluations done and results reported are convincing.

However it seem to me to lack a clear purpose, by attempting to be too many things at once and not quite succeeding properly at any of them. Is it an evaluation of standard deep learning architecture for nucleus segmentation in non trivial images? a report on active/interactive learning - human in the loop workflow to try and address the labeling issue for biological images, by introducing a new imagej plugin? or really the presentation of a new workflow of time course estimation from from fixed image ? According to the title (and indeed, in my humble opinion as well, since it is the most novel and innovative) it should be the later; however it is the least investigated and evaluated part, as opposed to DL, with one unique graph in the main paper and very little discussion. Suggestions could be:

- to study the influence of DL segmentation accuracy for time course estimation, providing evaluation of the workflow as a whole

- additional evaluation of the estimated time courses. Depending on the availability/feasibility of experiments, those could be done with simulation. At the very least corroboration of the found time courses from other experimental or biological arguments.

- application of the same technique on other published data

- provide some novel understanding or biological findings those data/methods allowed

- discuss some of the techniques that has been proposed for similar ends (https://doi.org/10.1038/nmeth.3545 or [https://doi.org/10.1038/s41467-017-00623-3](https://doi.org/10.1038/s41467-017-00623-3) comes to mind for example).

Other points:

- the DL methods are not compared with current/recent SOTA, like cellpose or stardist. Reporting their use (and potentially failure) would be interesting to the community

- The efficacy of deep learning for segmenting and classifying biological images is now well accepted and it may not be needed to insist on it to much

- The article talks about 'a small training dataset' but still many tens of images had to be manually segmented. Maybe a study of the influence of training dataset size on accuracy could be interesting, and/or a comparison of

- The assumption that "fluorescence intensity is proportional to protein concentration" could be discussed further, since it is not so obvious in IF. [30] seem to be about a novel technique for in vivo expression, it is not clear how it relates to that claim.

- Maybe a personal preference but bar plot + whiskers may not be the ideal way to report such results. For mean+-standard deviation, simple tables may suffice. Graphical representation could be more efficient when used to display whole distributions, like violin plots or similar techniques.

- more generally, unless I missed it, it seem the paper has been submitted as a research article and not specifically as a method paper as it is, it may lack novel biological findings for that, but would certainly have its place as a method paper, the comments above notwithstanding.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009949.r003

Decision Letter 1

Jason M Haugh, Attila Csikász-Nagy

8 Jan 2022

Dear Dr Pécot,

Thank you very much for submitting your manuscript "Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

The errors in the upload of Fig 2 and Fig 3 made it hard to follow the exact changes. Some of these have improved the manuscript. But several major points are still remains unanswered, thus another round of majr revision on the MS is needed. 

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Attila Csikász-Nagy

Associate Editor

PLOS Computational Biology

Jason Haugh

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: - I thank the authors for taking the time to review my feedback. I found tables S2 and S4 quite helpful to understand the different training pipelines. I would like to see a similar table when comparing the performance of the different methods: I would suggest a given IoU score (or, multiple) - and then comparing performance of all the different methods and showing those numbers in a table. Having to jump between tens of different subfigures to see the impact of data augmentation is very difficult and time consuming. I understand that a table cannot summarize all the results across the different IoU thresholds, and the authors can keep the figures in the supplementary material if necessary, but, it's really necessary to have a single table with the important performance metrics.

- Figure 2 and 3 are identical in my copy of the manuscript, and I could not evaluate that portion of the manuscript because I couldn't see the data.

- I still insist that if the authors would like to bin E2F data, they should show the raw histogram data justifying a binning procedure. They have that raw data (they need it to obtain the binned values) - and plotting a histogram, together with the cutoff thresholds should be very easy. I understand that this is standard in their field - but - for someone outside of the field looking to evaluate the manuscript, continuous data getting binned without seeing the raw distributions is confusing.

- The suggestions from reviewer 2 about presentation and multiple papers are great. Currently- this is two papers, and one paper might be better suited for a computer vision conference or workshop, while the biological findings and applications might belong here. For example - training curves and detailed comparisons to other methods would be really useful if this was a pure computer vision paper, but, in a biology paper, they are probably overkill, as people are not that interested in evaluating the methods at that level of detail.

Reviewer #2: The authors have taken into account most of the remarks made in previous version.

The manuscript is still very complicated to read, and it still lacks clarity of the objectives. However, I have no additional remark to bring and I consider it can be published as is.

Reviewer #3: The resubmitted article is much improved compared to the first version and in particular more focused. I have no major request of changes to be made before publication, just a few comments below.

The figures are wrong in the submitted manuscript (fig 2 and 3 are the same and Fig 4 seem unrelated). A bit problematic for the review but the main novelty are the simulations, which are in the supplementary. To be thoroughly checked before publication though...

Fig 3 has only the final results with all simulation and method detail relegated in the supplementary. I personally think it's a bit of a shame since the figures are the main points of entry for many reader, which would make that work invisible. I would welcome some of that in the main figures, but leave that choice to the authors/editors.

The simulations have pretty big error bars, and the noise seem to be only marginally affecting the results in many cases with averages being fairly constant from 0 to 30 and often 50% noise. Can the authors comment on that? Make one wonder if that is the most informative way of displaying that result, and what would the error bar on real data would be. Also the simulations are presented in the text before the time course estimation methods, while it uses/is meant to test it, which is a bit strange.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Attachment

Submitted filename: Figures_2_3.pdf

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009949.r005

Decision Letter 2

Jason M Haugh, Attila Csikász-Nagy

14 Feb 2022

Dear Dr Pécot,

Thank you very much for submitting your manuscript "Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

Please explain the method of thresholding, as requested by Ref. 1.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Attila Csikász-Nagy

Associate Editor

PLOS Computational Biology

Jason Haugh

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: I thank the authors for all the additions since the last review. I found the extra tables especially helpful, and appreciated the additional figure showing the E2F concentration.

In my view, the E2F quantization step is not justified on the basis of data: the distribution in figure S7 is obviously not multi-modal, and, in the material and methods section, the thresholding step was described as using 'manually defined thresholds', and there's no discussion at all about how those thresholds were calculated or how sensitive the results later in the paper are to the choice of this arbitrary threshold.

It might be that there is a biological justification for this thresholding step - and - at the very least, I'd like the authors discuss this in more detail in the main text. Currently, the main text has a quote:

"We assume that fluorescence intensity is proportional to protein =concentration [34, 35] and therefore define quantized levels of intensity for E2Fs."

Clearly, the second part (deciding to quantize a signal) does not follow from the first (protein concentration being proportional to the fluorescent signal).

Reviewer #3: All my comment on the text and figures have been addressed.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009949.r007

Decision Letter 3

Jason M Haugh, Attila Csikász-Nagy

21 Feb 2022

Dear Dr Pécot,

We are pleased to inform you that your manuscript 'Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Attila Csikász-Nagy

Associate Editor

PLOS Computational Biology

Jason Haugh

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009949.r008

Acceptance letter

Jason M Haugh, Attila Csikász-Nagy

8 Mar 2022

PCOMPBIOL-D-21-00751R3

Deep learning tools and modeling to estimate the temporal expression of cell cycle proteins from 2D still images

Dear Dr Pécot,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Zsofia Freund

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Example images.

    One example for each combination of markers and modality for fluorescence images used in the study. The combination of markers, number of images of each type, modality, and number of mice used in the study are shown on top of the images. Note the small size of the training set. Scale bar = 20μm.

    (TIF)

    S2 Fig. Evaluation of U-Net for nuclei segmentation.

    F1 score for multiple IoU thresholds obtained with the U-Net approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation or post-processing a-b, with data augmentation c-d, with data augmentation and corrected watershed postprocessing e-f. The lines correspond to the average F1 score while the areas represent the standard error.

    (TIF)

    S3 Fig. Evaluation of Inception-V3 for nuclear segmentation.

    F1 score for multiple IoU thresholds obtained with the Inception-V3 approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation or post-processing a-b, with data augmentation c-d, with data augmentation and corrected watershed postprocessing e-f. The lines correspond to the average F1 score while the areas represent the standard error.

    (TIF)

    S4 Fig. Evaluation of Mask R-CNN for nuclear segmentation.

    F1 score for multiple IoU thresholds obtained with the Mask R-CNN approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation but with transfer learning a-b, without transfer learning but with data augmentation c-d, with data augmentation and transfer learning e-f. The lines correspond to the average F1 score while the areas represent the standard error.

    (TIF)

    S5 Fig. Evaluation of Stardist for nuclear segmentation.

    F1 score for multiple IoU thresholds obtained with the Stardist approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation a-b and with data augmentation c-d. The lines correspond to the average F1 score while the areas represent the standard error.

    (TIF)

    S6 Fig. Evaluation of Cellpose for nuclear segmentation.

    F1 score for multiple IoU thresholds obtained with the Cellpose approach by trainig on confocal, half or all widefield images, and on confocal and widefield images without data augmentation a-b and with data augmentation c-d. The lines correspond to the average F1 score while the areas represent the standard error.

    (TIF)

    S7 Fig. Examples of intensity binning for E2F8.

    a-b E2F8 channel of a confocal a and a widefield b image. E2F8 positive nuclei are overlaid as orange circles while E2F8 negative nuclei are overlaid as red circles. Scale bar = 20μm. c-d Histograms of the E2F8 intensity observed in the nuclei shown in images a (c) and b (d). The 3 levels of intensity used for concentration estimation are displayed with different colors.

    (TIF)

    S8 Fig. Initialization of E2F3A and pH3 over time.

    a 2D histogram of E2F3A intensity and pH3 patterns. b Initialization of E2F3A and pH3 over time from the 2D histogram shown in a.

    (TIF)

    S9 Fig. Initialization of E2F3A and E2F4 over time.

    a 2D histogram of E2F3A and E2F4 intensity. b Modified 2D histogram of E2F3A and E2F4 intensity with vertex splitting and set of edges giving the minimum cost. c Initialization of E2F3A and E2F4 over time from the set of edges shown in b.

    (TIF)

    S10 Fig. Initialization of E2Fs with respect to EdU and pH3.

    a-b Initialization of E2F3A with respect to a EdU (n = 3) and b pH3 (n = 3). c-d Initialization of E2F8 concentration with respect to c EdU (n = 3) and d pH3 (n = 3). e-f Initialization of E2F4 concentration with respect to e EdU (n = 2) and f pH3 (n = 2). g-h Initialization of E2F3A concentration with respect to g E2F8 (n = 3) and h E2F4 (n = 2). For all curves, the E2Fs intensity for the different mice are represented as curves with different line styles while EdU and pH3 are shown above the E2Fs curves, with solid lines corresponding to diffuse states and dashed lines corresponding to punctate states.

    (TIF)

    S11 Fig. Initialization of E2Fs evolution over the cell cycle.

    a Temporal evolution of E2F3A over the cell cycle after registration between EdU and pH3 (n = 6). b Temporal evolution of E2F8 over the cell cycle after registration between EdU and pH3 (n = 6). c Temporal evolution of E2F4 over the cell cycle after registration between EdU and pH3 (n = 4). For all curves, the E2Fs intensity for the different mice are represented as curves with different line styles while EdU and pH3 are shown above the E2Fs curves, with solid lines corresponding to diffuse states and dashed lines corresponding to punctate states.

    (TIF)

    S12 Fig. Evaluation of the influence of the number of samples over the estimation of protein concentration over the cell cycle on simulated data.

    Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of samples when corrupting up to 50% of the simulated data with noise. n1 corresponds to the number of samples for E2F3A, E2F8, EdU and pH3 while n2 is the number of samples for E2F4. The lines correspond to the average mean squared error while the areas represent the standard error.

    (TIF)

    S13 Fig. Evaluation of the influence of the number of time bins over the estimation of protein concentration over the cell cycle on simulated data.

    Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of time bins when corrupting up to 50% of the simulated data with noise. The lines correspond to the average mean squared error while the areas represent the standard error.

    (TIF)

    S14 Fig. Evaluation of the influence of the number of intensity bins over the estimation of protein concentration over the cell cycle on simulated data.

    Mean squared error between the estimated and simulated concentrations of E2F3A a, E2F8 b, E2F4 c, EdU d and pH3 e by considering different numbers of intensity bins when corrupting up to 50% of the simulated data with noise. The lines correspond to the average mean squared error while the areas represent the standard error.

    (TIF)

    S1 Table. Time computation for training nuclei segmentation models.

    Computation time needed to train the five different deep learning approaches on the four training datasets with a GeForce RTX 2080 with Max-Q design. CF stands for confocal images, WF half stands for half the widefield images, WF stands for widefield images and CFWF stands for confocal and widefield images.

    (TIF)

    S2 Table. Training strategies for nuclei segmentation.

    Pre-processing, normalization, optimization method, data augmentation, transfer learning and post-processing used to train the five deep learning approaches.

    (TIF)

    S3 Table. Training data summary.

    Number of images and nuclei in the four different training and validation datasets. CF stands for confocal images, WF half stands for half the widefield images, WF stands for widefield images and CFWF stands for confocal and widefield images.

    (TIF)

    S4 Table. Time computation for processing nuclei segmentation.

    Computation time needed to process the five different deep learning approaches on the four training datasets with a GeForce RTX 2080 with Max-Q design.

    (TIF)

    S5 Table. Evaluation of deep learning approaches for nuclei segmentation in confocal images.

    F1 score obtained for IoU = 0.5 and IoU = 0.75 when segmenting nuclei in confocal images with U-Net, Inception-V3, Mask R-CNN, Stardist and Cellpose. DA stands for data augmentation, PP stands for post-processing, TL stands for transfer learning and IoU stands for intersection over union.

    (TIF)

    S6 Table. Evaluation of deep learning approaches for nuclei segmentation in widefield images.

    F1 score obtained for IoU = 0.5 and IoU = 0.75 when segmenting nuclei in widefield images with U-Net, Inception-V3, Mask R-CNN, Stardist and Cellpose. DA stands for data augmentation, PP stands for post-processing, TL stands for transfer learning and IoU stands for intersection over union.

    (TIF)

    S7 Table. Performance of Inception-V3 for marker identification with confocal images.

    Inception-V3 accuracy and standard error with and without transfer learning for marker identification of E2F3A, E2F8, EdU and pH3 in confocal images. DA stands for data augmentation.

    (TIF)

    S8 Table. Performance of Inception-V3 for marker identification with widefield images.

    Inception-V3 accuracy and standard error with and without transfer learning for marker identification of E2F3A, E2F8, E2F4, EdU and pH3 in widefield images. DA stands for data augmentation.

    (TIF)

    S9 Table. Time computation for training and processing marker identification.

    Computation time needed to train and process Inception-V3 for marker identification with a GeForce RTX 2080 with Max-Q design. DA stands for data augmentation.

    (TIF)

    Attachment

    Submitted filename: AnswersToReviewers.pdf

    Attachment

    Submitted filename: Figures_2_3.pdf

    Attachment

    Submitted filename: AnswersToReviewers.pdf

    Attachment

    Submitted filename: AnswersToReviewers.pdf

    Data Availability Statement

    Data availability All the images used in this study are available at https://data.mendeley.com/datasets/5r6kf37zd4/1. The training datasets for nuclei segmentation are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/UNet/datasets/nucleiSegmentation_E2Fs for the U-Net architecture and at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/MaskRCNN/datasets/nucleiSegmentation_E2Fs for the Mask R-CNN architecture. The training datasets for nuclei segmentation and marker identification are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification/tree/master/InceptionV3/trainingData for the Inception-V3 architecture. The images used for the evaluation and the ground truth are available at the same locations as the training datasets. The intensity 2D histograms used to estimate the E2Fs accumulation over the cell cycle are available at https://github.com/tpecot/EstimationOfProteinConcentrationOverTime/tree/master/data. Software availability The codes used to train and process deep learning approaches for nuclei segmentations and marker identification are available at https://github.com/tpecot/NucleiSegmentationAndMarkerIDentification. Archived code at time of publication: https://doi.org/10.5281/zenodo.4619243 [50] License: GPL3 The Octave code used to estimate the E2Fs accumulation over the cell cycle is available at https://github.com/tpecot/EstimationOfProteinConcentrationOverTime. Archived code at time of publication: https://doi.org/10.5281/zenodo.4639800 [51] License: GPL3 The Java code of the Annotater plugin and the plugin are available at https://github.com/tpecot/Annotater. Video tutorials to show how to use the Annotater are available at the same location. Archived code at time of publication: https://doi.org/10.5281/zenodo.4639802 [52] License: GPL3.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES