Abstract
Organoids have immense potential as ex vivo disease models for drug discovery and personalized drug screening. Dynamic changes in individual organoid morphology, number, and size can indicate important drug responses. However, these metrics are difficult and labor-intensive to obtain for high-throughput image datasets. Here, we present OrganoID, a robust image analysis platform that automatically recognizes, labels, and tracks single organoids, pixel-by-pixel, in brightfield and phase-contrast microscopy experiments. The platform was trained on images of pancreatic cancer organoids and validated on separate images of pancreatic, lung, colon, and adenoid cystic carcinoma organoids, which showed excellent agreement with manual measurements of organoid count (95%) and size (97%) without any parameter adjustments. Single-organoid tracking accuracy remained above 89% over a four-day time-lapse microscopy study. Automated single-organoid morphology analysis of a chemotherapy dose-response experiment identified strong dose effect sizes on organoid circularity, solidity, and eccentricity. OrganoID enables straightforward, detailed, and accurate image analysis to accelerate the use of organoids in high-throughput, data-intensive biomedical applications.
Author summary
A recent advance in biomedical research is the use of connective tissue gels to grow cells into microscopic structures, called organoids, that preserve and exhibit the physical and molecular traits of a particular organ. Organoids have enabled researchers to study complex phenomena, such as the beating of the heart or the folds of the intestines, and the effects of drugs on these properties. Changes in the size and shape of organoids are important indicators of drug response, but these are tedious to measure in large drug screening experiments, where thousands of microscopy images must be analyzed. We developed a software tool named OrganoID that automatically traces the exact shape of individual organoids in an image, even when multiple organoids are clumped together, and measures organoid changes over time. To show our tool in action, we used OrganoID to analyze pancreatic cancer organoids and their response to chemotherapy. We also showed that our tool can handle images of many different types of organoids, even those derived from mouse cells. With this software, researchers will be able to easily analyze immense quantities of organoid images in large-scale experiments to discover new drug treatments for a range of diseases.
This is a PLOS Computational Biology Software paper.
Introduction
Organoids are multicellular three-dimensional (3D) structures derived from primary or stem cells that are embedded into a biological matrix to create an extracellular environment that provides structural support and key growth factors. Organoids closely recapitulate cellular heterogeneity, structural morphology, and organ-specific functions of a variety of tissues, which improves modeling of in situ biological phenomena compared to traditional monolayer cell cultures [1]. Live-cell imaging can reveal important features of organoid dynamics, such as growth, apoptosis/necrosis, movement, and drug responses, which can reflect physiological and pathological events such as organ development, function, infarction, cancer, and infection [2–7].
While organoids have been successfully used to investigate important phenomena that might be obscured in simpler models, their use in data-intensive applications, such as high-throughput screening, has been more difficult. A major challenge for organoid experiments is drug-response measurement and analysis, which must be performed for a large number of microscopy images. Image analysis is particularly difficult for organoid experiments due to the movement of organoids across focal planes and variability in organoid size and shape between different tissue types, within the same tissue type, and within the same single culture sample [8,9]. Several recently developed organoid platforms, while powerful, rely on per-experiment or per-image tuning of brightfield image analysis parameters [10–12], or require manual labeling of each image [7,13,14], which limits experiment reproducibility and scale. Organoids can be fluorescently labeled to aid in image segmentation and tracking, such as through genetic modification for fluorescent protein expression [15–17], cell fixation and staining [10,16], or the use of membrane-permeable dyes. However, these approaches may alter intrinsic cellular dynamics from original samples [18,19], limit measurements to endpoint assays, or cause cumulative toxicity through longer growth times and limited diffusion through the hydrogel matrix [20]. There is a critical need for an automated image analysis tool that can robustly and reproducibly measure live-cell organoid responses in high-throughput experiments without the use of potentially toxic or confounding fluorescence techniques.
A number of software tools have been developed to automate the process of brightfield/phase-contrast organoid image analysis. These platforms use conventional image processing methods, such as adaptive thresholding and mathematical morphology [21], or convolutional neural networks [22–24] to identify organoids in sequences of microscopy images. Despite their advantages, many existing platforms require cellular nuclei to be transgenically labeled [23], which increases experiment time and complexity and may modify cellular dynamics. Other existing platforms require manual tuning of parameters for each image [21], focus on single-timepoint analysis [24], only provide population-averaged (bulk) measurements without single-organoid resolution, or are limited to bounding-box detection [22], which fails to capture potentially useful morphological information at the individual organoid level. Changes in organoid shape, such as spiking or blebbing, can reveal important responses to external stimuli and might be missed with bounding-box measurements [25]. Many of these existing platforms were also developed for the analysis of organoids derived from a single type of tissue and for images obtained with one specific optical configuration, precluding their use across different experimental settings.
To address these challenges, we have developed a software platform, OrganoID, that can identify and track individual organoids in a population derived from a wide range of tissue types, pixel by pixel, in both brightfield and phase-contrast microscopy images and in time-lapse experiments. OrganoID consists of (i) a convolutional neural network, which detects organoids in microscopy images, (ii) an identification module, which resolves contours to label individual organoids, and (iii) a tracking module, which follows identified organoids in time-lapse imaging experiments. Most importantly, OrganoID was able to accurately segment and track a wide range of organoid types, including those derived from pancreatic ductal adenocarcinoma, adenoid cystic carcinoma, and lung and colon epithelia, without cell labeling or parameter tuning. The OrganoID software overcomes a major hurdle to organoid image analysis and supports wider integration of the organoid model into high-throughput applications.
Design and implementation
A convolutional neural network for pixel-by-pixel organoid detection
We developed a deep learning-based image analysis pipeline, OrganoID, that recognizes and tracks individual organoids, pixel by pixel, for bulk and single-organoid analysis in brightfield and phase-contrast microscopy images (Fig 1A). The platform employs a convolutional neural network to transform microscopy images into organoid probability images, where brightness values represent the network belief that an organoid is present at a given pixel. The network structure was derived from the widely successful u-net approach to image segmentation [26] (Fig 1B). The u-net approach first passes each image through a contracting series of multidimensional convolutions and maximum filters to extract a set of deep feature maps that describe the image at various levels of detail and contexts, such as edges, shapes, and textures. The feature maps are then passed through an expanding series of transposed convolutions, which learn to localize the features and assemble a final output.
The OrganoID neural network follows the u-net structure and was optimized to require far fewer feature channels than the original implementation, as a reduction in trainable model parameters limits overfitting and minimizes the amount of memory and computational power required to use the network in the final distribution [27]. The original u-net implementation uses 64 filters in the first layer, which results in quite a large number of trainable parameters across the full network (over 30 million) [26]. We sequentially reduced the number of filters in the first layer by powers of two to reach a minimal value that preserved performance on the validation dataset (validation cross-entropy loss was minimal at 0.088 for 8 filters in the first layer, compared to 0.098 for 4 filters and 0.1 for 16 filters). The final OrganoID u-net structure uses only 8 filters in the first layer, which results in a network structure with less than 500,000 trainable parameters (a 98% reduction compared to the original implementation). All convolutional neurons were set to compute outputs with the exponential linear unit activation function, which has been previously shown to produce higher accuracy than the rectified linear unit (ReLU, used in the original u-net implementation) and avoid vanishing gradient problems during network training [28]. Neurons in the final convolutional layer were set with a sigmoid activation function to produce a normalized output that corresponds to the probability that an organoid is present at each pixel in the original image. All images are automatically contrasted and resized to 512x512 pixels before training and inference. Python was used for the entire OrganoID platform and Keras (a neural network interface for the TensorFlow software library) was used for network expression, training, and operation.
An original dataset of 66 brightfield and phase-contrast microscopy images of organoids were manually segmented to produce black-and-white ground truth images for network training (52 image pairs) and validation (14 image pairs) (S1A Fig). Each image featured 5 to 50 organoids derived from human pancreatic ductal adenocarcinoma (PDAC) samples from two different patients. Organoids were either grown on a standard tissue culture plate or on a microfluidic organoid platform [29]. To teach the network that segmentations should be independent of imaging orientation, field of view, lens distortion, and other potential sources of overfitting, the training dataset was processed with the Augmentor Python package [30] with random rotation, zoom, elastic distortion, and shear transformations to produce an augmented set of 2,000 images for training (S1B Fig). The neural network was trained on this dataset with the Adam stochastic optimization algorithm [31] at a learning rate of 0.001. Unweighted binary cross-entropy between predicted and ground-truth segmentations was used for the loss function. Layer weights were initialized with the He method [32]. An early stopping rule was used to halt training once performance on the validation dataset reached a minimum (i.e. once 10 epochs pass with no improvement in validation loss). The batch size was set to train on 8 augmented images for every round of backpropagation. Dropout regularization was introduced after all convolutions to randomly set 12.5% of neuron weights to zero after each batch. After each epoch, a copy of the model was saved for additional evaluation of the training process.
Identification of individual organoids with diverse morphology and size
The convolutional neural network detects organoids in an image on a pixel-by-pixel basis, which can be used to measure bulk responses. For single-organoid analysis, pixels must be grouped together to identify individual organoids. This task is straightforward for isolated organoids, where all high-belief pixels in a cluster correspond to one organoid, but is more difficult for organoids that are in physical contact. To address this challenge, we developed an organoid separation pipeline (Fig 2A) that uses the raw network prediction image to group pixels into single-organoid clusters. Conventionally, neural network image segmentation methods set an absolute threshold on predicted pixels to produce a binary detection mask. This approach is effective but discards useful information about the strength of predictions. The segmentations that were used to train the neural network were produced with a 2-pixel boundary between organoids in contact that follows the contour of separation. As a result, the network predictions were marginally less confident about pixels near organoid boundaries (Fig 2A). We took advantage of this phenomenon to identify organoid contours with a Canny edge detector [33], an image transformation which (i) computes the partial derivative of pixel intensity to identify high-contrast areas, (ii) blurs the image to smooth noisy regions, and (iii) applies a hysteresis-based threshold to identify locally strong edges. Edges are removed from the thresholded prediction image to mark the centers of each organoid. These centers are then used as initializer basins in a watershed transformation [34], an algorithm inspired by the filling of geological drainage basins that is used to segment contacting objects in an image. The image is further refined to remove organoids that may be partially out of the field-of-view or below a particular size threshold. The pipeline outputs a labeled image, where the pixels that represent an individual organoid are all set to a unique organoid ID number, which can be used for single-organoid analysis (Fig 2A).
The scipy [35] and scikit-image [36] packages were used to identify individual organoids from the network detection images. Detection images were thresholded at 0.5 and passed through a morphological opening operation to remove weak and noisy predictions. Partial derivatives were computed with a Sobel filter, passed through a Gaussian filter (σ = 2), and converted to an edge mask with a hysteresis threshold filter (Thi = 0.05, Tlo = 0.005), which was then was used to mark organoid centers. The watershed method was used to identify filled organoid contours, with the organoid centers as label initializers and the network prediction image as an inverted heightmap. Labeled organoids were morphologically filled and discarded if the total area was below 200 pixels. Images were morphologically processed with scikit-image to record properties of identified organoids.
Automated organoid tracking in time-lapse microscopy experiments
OrganoID was also developed with a tracking module for longitudinal single-organoid analysis of time-course imaging experiments, where changes in various properties of individual organoids, such as size and shape, can be measured and followed over time. The central challenge for the tracking module is to match a detected organoid in a given image to the same organoid in a later image. The OrganoID tracking algorithm builds a cost-assignment matrix for each image in a time-lapse sequence, where each row corresponds to an organoid tracked from a previous image and each column corresponds to an organoid detected in the current image. Each matrix entry is the cost of assigning a given organoid detection in the current image to a detection in the previous image. The cost function was defined as the inverse of the number of shared pixels between the two organoids. As such, the assignment cost between two detections will be minimal for those that are of similar shape and in a similar location. The cost-assignment matrix is also padded with additional rows and columns to allow for “pseudo-assignments” that represent missing or newly detected organoids. Finally, the Munkres variant of the Hungarian algorithm [37] was used to minimize the cost-assignment matrix to find an optimal matching between organoid detections in the previous image and the detections in the current image. This approach produces organoid “tracks” for unique organoids in time-lapse images.
Results
Network training halted after 37 epochs, when segmentation error (binary cross-entropy) on the validation dataset converged to a minimal value (S1C Fig and S1 Video). After hyperparameter tuning, the final model performance was assessed on a novel PDAC organoid testing dataset, previously unseen by the network. Performance was quantified with the intersection-over-union (IOU) metric, which is defined as the overlap between the predicted and actual organoid pixels divided by their union (Fig 1C). An IOU greater than 0.5 is generally considered to reflect a good prediction [38,39] and we chose this value to be the minimal benchmark for satisfactory model performance. All PDAC images passed the benchmark, with a mean IOU of 0.74 (SD = 0.081) (S2 Fig).
Because the network was trained and assessed solely with images of PDAC organoids, we were also curious to evaluate its capacity to generalize to organoids derived from non-PDAC tissues. Microscopy images of organoids derived from lung epithelia, colon epithelia, and salivary adenoid cystic carcinoma (ACC) were manually segmented (6 images for each tissue type). The OrganoID network passed the benchmark for all of these non-PDAC images, with a mean IOU of 0.79 (SD = 0.096) (S2 Fig). Additionally, 19 images of organoids derived from mouse small intestine were segmented and evaluated with the model and passed the benchmark with a mean IOU of 0.70 (SD = 0.15). These results support the potential of the OrganoID neural network to generalize to organoids from various tissues of origin without parameter tuning (Fig 1D). However, 3 images of mouse small intestinal organoids did not pass the benchmark, which suggests a limit to this generalizability. To demonstrate how performance can be improved on novel datasets, a copy of the neural network was additionally trained on a subset of mouse small intestinal organoid images. This retrained network showed considerable improvement on a test dataset of mouse intestinal organoids with a mean IOU of 0.82 (SD = 0.09) while maintaining performance on the original test dataset with a mean IOU of 0.76 (SD = 0.10) (S2 Fig).
The network was also evaluated for appropriate exclusion of non-organoid technical artifacts. Air bubbles in culture media or gel matrix were rarely detected by OrganoID with a false positive rate of 4.2% (S3A Fig). We observed that OrganoID segmentations also avoided cellular debris or dust embedded into the gel, ignored chamber borders, and performed robustly across microscope resolutions, organoid concentrations, and organoid shapes (S3B–S3G Fig). OrganoID proved to be computationally efficient; each image was segmented in ~300 milliseconds on a laptop CPU (Intel i7-9750H, 2.6GHz) with less than 200 megabytes of RAM usage.
For quantitative validation of single-organoid analysis, OrganoID was used to count and measure the area of organoids in images from the PDAC and non-PDAC human testing datasets (a total of 28 images). These data were then compared to the number and area of organoids in the corresponding manually segmented images. Organoid counts agreed with a concordance correlation coefficient (CCC) of 0.95 [95% CI 0.89–0.98]. OrganoID, on average, detected 1.1 fewer organoids per image than manual segmentation. The limits of agreement between OrganoID and manual counts were between -10.35 and 8.14 organoids (Fig 2B). Organoid area comparison demonstrated a CCC of 0.97 [95% CI 0.96–0.97]. OrganoID area measurements were biased to be 1.06x103 μm2 larger per organoid with limits of agreement between -3.42x103 μm2 and 5.55x103 μm2 (Fig 2C). Several additional morphological metrics were calculated, including circularity (the ratio of the organoid perimeter to the perimeter of a perfect circle), eccentricity (a measure of elliptical deviation from a circle), and solidity (the ratio of the organoid area to the area of its convex hull). These three metrics showed moderate concordance (0.52, 0.41, and 0.59 respectively) between automated and manual measurements (S4 Fig). Performance on brightfield microscopy images was higher than performance on phase-contrast microscopy images. Overall, the measurements produced by OrganoID were in considerable agreement with those obtained by hand in our test dataset, which supports the use of OrganoID for automated single-organoid analysis.
For validation of the organoid tracking module, microscopy images taken every 2 hours from a 92-hour organoid culture experiment were passed through the entire OrganoID pipeline to produce growth curves that followed single-organoid changes over time (Fig 2D–2E). The tracking step was also performed by hand to evaluate automated performance. OrganoID tracking maintained over 89% accuracy (defined as the fraction of identified organoids that were correctly matched at each time step) throughout the duration of the experiment (S5 Fig and S2 Video). Tracking accuracy was not affected when frames were discarded to simulate a 12-hour imaging interval, which demonstrates low susceptibility to imaging frequency.
OrganoID normalizes single-organoid death responses over time
PDAC organoids were treated with serial dilutions of gemcitabine (3nM-1000nM), an FDA-approved chemotherapeutic agent commonly used to treat pancreatic cancer. Propidium iodide (PI), a fluorescent reporter of cellular necrosis, was also added to the culture media to monitor death responses. Cultures were then imaged every 4 hours over 72 hours and the produced brightfield images were processed with the OrganoID platform to identify organoids and analyze timelapse features (Fig 3A). At gemcitabine concentrations above 3 nM, the total organoid area increased for the first several hours, which reflected initial organoid growth, but then decreased to a value and at a rate inversely proportional to gemcitabine concentration (Fig 3B). Identified organoid counts for gemcitabine concentrations above 10 nM also sharply decreased over time. The total fluorescence intensity of the PI signal increased over time to a value and at a rate proportional to gemcitabine concentration, however the response to 100 nM gemcitabine appeared to induce a stronger death signal than the response to 1000 nM gemcitabine (Fig 3C).
Measurements of death stains such as PI are typically normalized to a viability measurement that accounts for differences in the number and size of organoids between replicates that can compound over the duration of an experiment. We hypothesized that OrganoID could be used to measure morphological features indicative of organoid viability over time, which could then be used to normalize and decrease the variability of measurements across experimental replicates. Furthermore, organoid tracking with OrganoID would enable per-organoid measurement normalization, which we observed to significantly decrease the coefficient of variation of organoid area and fluorescence change across organoids imaged at the same time point and exposed to the same concentration of gemcitabine (S6 Fig). Normalization of PI intensity with tracked single-organoid area indeed increased the separation of responses between each treatment group over time, decreased standard error across replicates, and corrected the response discrepancy between the 100 nM and 1000 nM conditions (Fig 3C). We also compared dose-response measurements to an MTS proliferation assay ((3-(4,5-dimethylthiazol-2-yl)-5-(3-carboxymethoxyphenyl)-2-(4-sulfophenyl)-2H-tetrazolium), a gold-standard indicator of cell viability, at the experiment endpoint. The MTS assay determined the half-maximal effective concentration (EC50) to be 26.13 nM of gemcitabine. However, total fluorescence and organoid area underestimated this value to be 16.95 nM and 12.67 nM, respectively. In contrast, the dose response of area-normalized fluorescence determined the EC50 to be 24.45 nM, considerably closer to the MTS standard (Fig 3D).
OrganoID uncovers morphology changes predictive of drug response
Changes in organoid morphology can indicate important phenotypic responses and state transitions. For example, some tumor organoid models grow into structures with invasive projections into the culture matrix, reflecting epithelial-mesenchymal transition; the addition of certain chemotherapeutic agents prevents development of these protrusions with minimal effects on overall organoid size [25]. These important responses can be investigated through single-organoid image analysis that captures the precise contour of each organoid. We used OrganoID to automatically profile the morphology of individual organoids across the gemcitabine dose-response experiment. Organoid circularity, solidity, and eccentricity showed sigmoidal dose responses and determined the endpoint EC50 to be 28.47 nM, 27.05 nM, and 18.08 nM, respectively (S7 Fig). The effect size of gemcitabine concentration on these shape metrics was also computed with Cohen’s d statistic, defined as the ratio of the difference in means to the pooled standard deviation. Gemcitabine dosage had a moderate to large endpoint effect on organoid circularity, solidity, and eccentricity across concentrations below and above the EC50 (Fig 3E). Circularity and solidity decreased at higher concentrations, while eccentricity increased, reflecting disrupted organoid morphology. This worked example demonstrates the advantages of OrganoID for automated bulk and single-organoid morphological analysis of time-course experiments without the need for live-cell fluorescence techniques.
Availability and future directions
Organoids have revolutionized biomedical research through improved model representation of native tissues and organ systems. However, the field has yet to fully enter the high-throughput experimental space. A central bottleneck is the challenge of automated response measurement and analysis in large numbers of microscopy images. Organoids exhibit striking diversity in morphology and size and can move through their 3D environment into and out of the focal plane; current image processing tools have not quite been able to capture these aspects in a robust manner. We developed OrganoID to bridge this gap and automate the process of accurate pixel-by-pixel organoid identification and tracking over space and time.
Experimental replicates in organoid studies can differ in the number and distribution of sizes of organoids. This difference between identical conditions requires per-sample normalization of response measurements to baseline growth of organoid colonies. There are several commercially available live-cell assays that can facilitate normalization in 2D culture. However, these same assays have proven difficult for organoid use due to the production of toxic photobleaching byproducts, limited diffusion through the gel matrix, and nonspecific staining of the gel matrix that results in a considerable background signal. Another available option is to genetically modify each organoid sample to express fluorescent proteins; however, this method increases experimental time and complexity and may alter cellular dynamics from the original tissue sample. The OrganoID platform can be leveraged for accurate normalization of standard organoid assays without live-cell fluorescence methods. OrganoID is also uniquely useful for efficient quantification of single-organoid morphological features, such as circularity, solidity, and eccentricity, that can reflect important dynamic responses. In our gemcitabine dose-response experiment, we observed that PDAC organoid circularity and solidity decreased with an increase in gemcitabine dosage. This disturbance of organoid architecture is likely due to the interference of gemcitabine with RNA and DNA synthesis [40], which may in turn affect cell turnover and production of signaling and structural proteins.
In this work, we have also contributed a manually segmented organoid image dataset for use in other computational platforms. OrganoID has demonstrated compatibility with organoids of various sizes, shapes, and sample concentrations as well as various optical configurations. Most excitingly, the OrganoID model was trained and validated on images of PDAC organoids but still demonstrated excellent generalization to images of other types of organoids, including those derived from colon tissue, lung tissue, and adenoid cystic carcinoma.
OrganoID was trained with and tested on a diverse yet relatively small set of images. Despite the suggested generalizability of OrganoID to various samples and optical configurations, performance may still differ with other types of organoids with significantly different morphology. Performance on a dataset of mouse small intestinal organoids passed our benchmark on average, however there were several challenging images that were not processed adequately. We demonstrated further training on a handful of images of mouse small intestinal organoids to improve performance on such challenging datasets. As well, OrganoID can only detect and assign a single organoid to each pixel in an image. While the platform can appropriately identify contours of organoids in physical contact, it cannot distinguish organoids that overlap across the focal plane. These limitations could be overcome in future work with a network model that produces multiple outputs per pixel as well as additional validation, an expanded training dataset, and the use of multiple focal planes for image analysis.
We have released the OrganoID platform open-source and freely licensed on GitHub (https://github.com/jono-m/OrganoID). The repository includes all source code as well as usage instructions and scripts used for the examples presented in this paper. The network training module is also included on the repository to allow further model training to improve performance for any untested applications. The training and testing dataset is available at through the Open Science Framework (https://osf.io/xmes4/). Our image analysis platform serves as an important tool for the use of organoids as physiologically relevant models in high-throughput research. The platform can accurately capture detailed morphological measurements of individual organoids in live-cell microscopy experiments without the use of genetic modifications or potentially cytotoxic dyes. These metrics can reveal important organoid responses that might be otherwise obscured with the use of bounding-box tools. The ability of the platform to generalize to a range of organoid types without parameter tuning also reflects the potential for the platform to standardize morphological assay readouts and improve measurement reproducibility. OrganoID achieves comprehensive and expedient image analysis of organoid experiments to enable the broader use of organoids as tissue models for high-throughput investigations into biological systems.
Supporting information
Data Availability
We have released the OrganoID platform open-source and freely licensed on GitHub (https://github.com/jono-m/OrganoID). The repository includes all source code as well as the entire training and testing dataset, usage instructions, and scripts used for the examples presented in this paper. The network training module is also included on the repository to allow further model training to improve performance for any untested applications. The dataset is also available at https://osf.io/xmes4/.
Funding Statement
This work is supported by NIH R01 GM127527 and P. G. Allen Distinguished Investigator Award (https://pgafamilyfoundation.org/) to S.T. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Kretzschmar K, Clevers H. Organoids: Modeling Development and the Stem Cell Niche in a Dish. Developmental Cell [Internet]. 2016. Sep;38(6):590–600. Available from: doi: 10.1016/j.devcel.2016.08.014 [DOI] [PubMed] [Google Scholar]
- 2.Driehuis E, Kolders S, Spelier S, Lõhmussaar K, Willems SM, Devriese LA, et al. Oral Mucosal Organoids as a Potential Platform for Personalized Cancer Therapy. Cancer Discovery [Internet]. 2019. May 3;9(7):852–71. Available from: 10.1158/2159-8290.cd-18-1522. [DOI] [PubMed] [Google Scholar]
- 3.Dutta D, Heo I, Clevers H. Disease Modeling in Stem Cell-Derived 3D Organoid Systems. Trends in Molecular Medicine [Internet]. 2017. May;23(5):393–410. Available from: doi: 10.1016/j.molmed.2017.02.007 [DOI] [PubMed] [Google Scholar]
- 4.Sachs N, de Ligt J, Kopper O, Gogola E, Bounova G, Weeber F, et al. A Living Biobank of Breast Cancer Organoids Captures Disease Heterogeneity. Cell [Internet]. 2018. Jan;172(1–2):373–386.e10. Available from: doi: 10.1016/j.cell.2017.11.010 [DOI] [PubMed] [Google Scholar]
- 5.Clevers H, Tuveson DA. Organoid Models for Cancer Research. Annual Review of Cancer Biology [Internet]. 2019. Mar 4;3(1):223–34. Available from: doi: 10.1146/annurev-cancerbio-030518-055702 [DOI] [Google Scholar]
- 6.van der Vaart J, Böttinger L, Geurts MH, van de Wetering WJ, Knoops K, Sachs N, et al. Modelling of primary ciliary dyskinesia using patient-derived airway organoids. EMBO Rep [Internet]. 2021. Dec 6 [cited 2022 Aug 8];22(12). Available from: https://onlinelibrary.wiley.com/doi/10.15252/embr.202052058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Richards DJ, Li Y, Kerr CM, Yao J, Beeson GC, Coyle RC, et al. Human cardiac organoids for the modelling of myocardial infarction and drug cardiotoxicity. Nat Biomed Eng [Internet]. 2020. Apr [cited 2022 Apr 8];4(4):446–62. Available from: http://www.nature.com/articles/s41551-020-0539-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Garreta E, Kamm RD, Chuva de Sousa Lopes SM, Lancaster MA, Weiss R, Trepat X, et al. Rethinking organoid technology through bioengineering. Nature Materials [Internet]. 2020. Nov 16;20(2):145–55. Available from: doi: 10.1038/s41563-020-00804-4 [DOI] [PubMed] [Google Scholar]
- 9.Pettinato G, Coughlan MF, Zhang X, Chen L, Khan U, Glyavina M, et al. Spectroscopic label-free microscopy of changes in live cell chromatin and biochemical composition in transplantable organoids. Sci Adv [Internet]. 2021. Aug 18;7(34):eabj2800. Available from: https://pubmed.ncbi.nlm.nih.gov/34407934. doi: 10.1126/sciadv.abj2800 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Mead BE, Hattori K, Levy L, Imada S, Goto N, Vukovic M, et al. Screening for modulators of the cellular composition of gut epithelia via organoid models of intestinal stem cell differentiation. Nature Biomedical Engineering [Internet]. 2022. Mar 21; Available from: doi: 10.1038/s41551-022-00863-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Brandenberg N, Hoehnel S, Kuttler F, Homicsko K, Ceroni C, Ringel T, et al. High-throughput automated organoid culture via stem-cell aggregation in microcavity arrays. Nat Biomed Eng [Internet]. 2020. Sep [cited 2022 Apr 8];4(9):863–74. Available from: http://www.nature.com/articles/s41551-020-0565-2. doi: 10.1038/s41551-020-0565-2 [DOI] [PubMed] [Google Scholar]
- 12.Zhou Z, Van der Jeught K, Fang Y, Yu T, Li Y, Ao Z, et al. An organoid-based screen for epigenetic inhibitors that stimulate antigen presentation and potentiate T-cell-mediated cytotoxicity. Nat Biomed Eng [Internet]. 2021. Nov [cited 2022 Apr 8];5(11):1320–35. Available from: https://www.nature.com/articles/s41551-021-00805-x. doi: 10.1038/s41551-021-00805-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Wiedenmann S, Breunig M, Merkle J, von Toerne C, Georgiev T, Moussus M, et al. Single-cell-resolved differentiation of human induced pluripotent stem cells into pancreatic duct-like organoids on a microwell chip. Nat Biomed Eng [Internet]. 2021. Aug [cited 2022 Apr 8];5(8):897–913. Available from: https://www.nature.com/articles/s41551-021-00757-2. doi: 10.1038/s41551-021-00757-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Tanaka N, Kanatani S, Kaczynska D, Fukumoto K, Louhivuori L, Mizutani T, et al. Three-dimensional single-cell imaging for the analysis of RNA and protein expression in intact tumour biopsies. Nat Biomed Eng [Internet]. 2020. Sep [cited 2022 Apr 8];4(9):875–88. Available from: http://www.nature.com/articles/s41551-020-0576-z. doi: 10.1038/s41551-020-0576-z [DOI] [PubMed] [Google Scholar]
- 15.Kim S, Choung S, Sun RX, Ung N, Hashemi N, Fong EJ, et al. Comparison of Cell and Organoid-Level Analysis of Patient-Derived 3D Organoids to Evaluate Tumor Cell Growth Dynamics and Drug Response. SLAS Discov [Internet]. 2020/04/30 ed. 2020. Aug;25(7):744–54. Available from: https://pubmed.ncbi.nlm.nih.gov/32349587. doi: 10.1177/2472555220915827 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Dekkers JF, Alieva M, Wellens LM, Ariese HCR, Jamieson PR, Vonk AM, et al. High-resolution 3D imaging of fixed and cleared organoids. Nature Protocols [Internet]. 2019. May 3;14(6):1756–71. Available from: doi: 10.1038/s41596-019-0160-8 [DOI] [PubMed] [Google Scholar]
- 17.Hof L, Moreth T, Koch M, Liebisch T, Kurtz M, Tarnick J, et al. Long-term live imaging and multiscale analysis identify heterogeneity and core principles of epithelial organoid morphogenesis. BMC Biol [Internet]. 2021. Feb 24;19(1):37–37. Available from: https://pubmed.ncbi.nlm.nih.gov/33627108. doi: 10.1186/s12915-021-00958-w [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Bailey SR, Maus MV. Gene editing for immune cell therapies. Nature Biotechnology [Internet]. 2019. Jun 3;37(12):1425–34. Available from: 10.1038/s41587-019-0137-8. [DOI] [PubMed] [Google Scholar]
- 19.Ang LT, Tan AKY, Autio MI, Goh SH, Choo SH, Lee KL, et al. A Roadmap for Human Liver Differentiation from Pluripotent Stem Cells. Cell Rep [Internet]. 2018. Feb 20;22(8):2190–205. Available from: https://pubmed.ncbi.nlm.nih.gov/29466743. doi: 10.1016/j.celrep.2018.01.087 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Riss T, Trask OJ Jr. Factors to consider when interrogating 3D culture models with plate readers or automated microscopes. In Vitro Cell Dev Biol Anim [Internet]. 2021/02/09 ed. 2021. Feb;57(2):238–56. Available from: https://pubmed.ncbi.nlm.nih.gov/33564998. doi: 10.1007/s11626-020-00537-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Borten MA, Bajikar SS, Sasaki N, Clevers H, Janes KA. Automated brightfield morphometry of 3D organoid populations by OrganoSeg. Sci Rep [Internet]. 2018. Mar 28;8(1):5319–5319. Available from: https://pubmed.ncbi.nlm.nih.gov/29593296. doi: 10.1038/s41598-017-18815-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Kassis T, Hernandez-Gordillo V, Langer R, Griffith LG. OrgaQuant: Human Intestinal Organoid Localization and Quantification Using Deep Convolutional Neural Networks. Sci Rep [Internet]. 2019. Aug 28;9(1):12479–12479. Available from: https://pubmed.ncbi.nlm.nih.gov/31462669. doi: 10.1038/s41598-019-48874-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Kok RNU, Hebert L, Huelsz-Prince G, Goos YJ, Zheng X, Bozek K, et al. OrganoidTracker: Efficient cell tracking using machine learning and manual error correction. PLoS One [Internet]. 2020. Oct 22;15(10):e0240802–e0240802. Available from: https://pubmed.ncbi.nlm.nih.gov/33091031. doi: 10.1371/journal.pone.0240802 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Larsen BM, Kannan M, Langer LF, Leibowitz BD, Bentaieb A, Cancino A, et al. A pan-cancer organoid platform for precision medicine. Cell Reports [Internet]. 2021. Jul [cited 2022 Apr 8];36(4):109429. Available from: https://linkinghub.elsevier.com/retrieve/pii/S2211124721008469. doi: 10.1016/j.celrep.2021.109429 [DOI] [PubMed] [Google Scholar]
- 25.Zhao N, Powell RT, Yuan X, Bae G, Roarty KP, Stossi F, et al. Morphological screening of mesenchymal mammary tumor organoids to identify drugs that reverse epithelial-mesenchymal transition. Nat Commun [Internet]. 2021. Jul 12;12(1):4262–4262. Available from: https://pubmed.ncbi.nlm.nih.gov/34253738. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. Lecture Notes in Computer Science [Internet]. 2015;234–41. Available from: 10.1007/978-3-319-24574-4_28. [DOI] [Google Scholar]
- 27.Bejani MM, Ghatee M. A systematic review on overfitting control in shallow and deep neural networks. Artif Intell Rev [Internet]. 2021. Dec [cited 2022 Aug 8];54(8):6391–438. Available from: https://link.springer.com/10.1007/s10462-021-09975-1. [Google Scholar]
- 28.Clevert DA, Unterthiner T, Hochreiter S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). 2015. [cited 2022 Apr 8]; Available from: https://arxiv.org/abs/1511.07289. [Google Scholar]
- 29.Schuster B, Junkin M, Kashaf SS, Romero-Calvo I, Kirby K, Matthews J, et al. Automated microfluidic platform for dynamic and combinatorial drug screening of tumor organoids. Nat Commun [Internet]. 2020. Oct 19;11(1):5271–5271. Available from: https://pubmed.ncbi.nlm.nih.gov/33077832. doi: 10.1038/s41467-020-19058-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.D Bloice M, Stocker C, Holzinger A. Augmentor: An Image Augmentation Library for Machine Learning. The Journal of Open Source Software [Internet]. 2017. Nov 4;2(19):432. Available from: 10.21105/joss.00432. [DOI] [Google Scholar]
- 31.Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. 2014. [cited 2022 Apr 8]; Available from: https://arxiv.org/abs/1412.6980. [Google Scholar]
- 32.He K, Zhang X, Ren S, Sun J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In IEEE; 2015. Available from: doi: 10.1109/iccv.2015.123 [DOI] [Google Scholar]
- 33.Canny J. A Computational Approach to Edge Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence [Internet]. 1986. Nov;PAMI-8(6):679–98. Available from: 10.1109/tpami.1986.4767851. [DOI] [PubMed] [Google Scholar]
- 34.Roerdink JBTM, Meijster A. The Watershed Transform: Definitions, Algorithms and Parallelization Strategies. Fundamenta Informaticae [Internet]. 2000. [cited 2022 Aug 2];41(1,2):187–228. Available from: https://www.medra.org/servlet/aliasResolver?alias=iospress&doi=10.3233/FI-2000-411207. [Google Scholar]
- 35.Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat Methods [Internet]. 2020. Mar 2 [cited 2022 Apr 8];17(3):261–72. Available from: http://www.nature.com/articles/s41592-019-0686-2. doi: 10.1038/s41592-019-0686-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.van der Walt S, Schönberger JL, Nunez-Iglesias J, Boulogne F, Warner JD, Yager N, et al. scikit-image: image processing in Python. PeerJ [Internet]. 2014. Jun 19;2:e453–e453. Available from: https://pubmed.ncbi.nlm.nih.gov/25024921. doi: 10.7717/peerj.453 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Munkres J. Algorithms for the Assignment and Transportation Problems. Journal of the Society for Industrial and Applied Mathematics [Internet]. 1957. Mar;5(1):32–8. Available from: doi: 10.1137/0105003 [DOI] [Google Scholar]
- 38.Englbrecht F, Ruider IE, Bausch AR. Automatic image annotation for fluorescent cell nuclei segmentation. Raja G, editor. PLoS ONE [Internet]. 2021. Apr 16 [cited 2022 Aug 2];16(4):e0250093. Available from: https://dx.plos.org/10.1371/journal.pone.0250093. doi: 10.1371/journal.pone.0250093 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Long F. Microscopy cell nuclei segmentation with enhanced U-Net. BMC Bioinformatics [Internet]. 2020. Dec [cited 2022 Aug 2];21(1):8. Available from: https://bmcbioinformatics.biomedcentral.com/articles/10.1186/s12859-019-3332-1. doi: 10.1186/s12859-019-3332-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Mini E, Nobili S, Caciagli B, Landini I, Mazzei T. Cellular pharmacology of gemcitabine. Annals of Oncology [Internet]. 2006. May [cited 2022 Aug 8];17:v7–12. Available from: https://linkinghub.elsevier.com/retrieve/pii/S0923753419381785. doi: 10.1093/annonc/mdj941 [DOI] [PubMed] [Google Scholar]