Skip to main content
Frontiers in Bioinformatics logoLink to Frontiers in Bioinformatics
. 2023 Jul 26;3:1243663. doi: 10.3389/fbinf.2023.1243663

Digital staining facilitates biomedical microscopy

Michael John Fanous 1, Nir Pillar 1,2, Aydogan Ozcan 1,2,3,4,*,
PMCID: PMC10411189  PMID: 37564725

Abstract

Traditional staining of biological specimens for microscopic imaging entails time-consuming, laborious, and costly procedures, in addition to producing inconsistent labeling and causing irreversible sample damage. In recent years, computational “virtual” staining using deep learning techniques has evolved into a robust and comprehensive application for streamlining the staining process without typical histochemical staining-related drawbacks. Such virtual staining techniques can also be combined with neural networks designed to correct various microscopy aberrations, such as out-of-focus or motion blur artifacts, and improve upon diffracted-limited resolution. Here, we highlight how such methods lead to a host of new opportunities that can significantly improve both sample preparation and imaging in biomedical microscopy.

Keywords: biomedical microscopy, computational imaging, computational staining, digital staining, virtual staining, quantitative phase imaging, intelligent microscopy, digital pathology

Introduction

Histochemical staining is an integral part of well-established pathology clinical workflows. Since thin tissue sections are mostly transparent, their features cannot be adequately observed through a standard brightfield microscope without exogenous chromatic staining. Another exogenous label commonly used to study biological specimens is formed by fluorescent probes, which enable highly specific tracking of sample components (Lichtman and Conchello, 2005) and can be used to monitor, e.g., nuclear dynamics (Kandel et al., 2020) and cellular viability (Hu et al., 2022). However, these labeling processes are time-consuming and laborious, comprising sample fixation, embedding, sectioning, and staining (Alturkistani et al., 2016). Furthermore, staining is not a perfectly repeatable procedure considering variations among human operators/technicians, and therefore the exact distribution and intensity of stains may differ from one staining operation to the next. Another disadvantage of exogenous staining is, in general, associated with their destructive nature as well as phototoxicity and photobleaching (Ounkomol et al., 2018; Jo et al., 2021; He et al., 2022), limiting imaging durations and compromising the integrity of the samples and their labels over time. Moreover, these staining procedures introduce distortions to the tissue that prevent further labelling or molecular analysis on the same regions, which presents a significant limitation in cases where multiple stains are required (Pillar and Ozcan, 2022).

An alternative approach to measuring the features of transparent biological samples is to exploit their inherent optical properties, such as autofluorescence or optical path length, in order to generate a contrast of their constituents. Autofluorescence (Monici, 2005), phase-contrast (Burch and Stock, 1942) and differential interference contrast (DIC) (Lang, 1982) microscopy offer such label-free information. Quantitative phase imaging (QPI) (Majeed et al., 2017; Park et al., 2018; Majeed et al., 2019) techniques, which provide precise phase data at the pixel level, have proven especially useful in biological applications.

Two general obstacles to using these modalities for a wider range of biomedical applications include: 1) most pathologists or medical experts have no familiarity with these kinds of images and cannot effectively interpret them, and 2) they lack the subcellular and molecular specificity that extraneous labels provide.

In recent years, largely due to the extraordinary progress in machine learning capabilities, computational staining techniques have emerged as an elegant solution to overcome these issues. Deep learning networks have been built to derive the stain of interest synthetically—whether chromatic or fluorescent—from label-free images (Rivenson et al., 2019a; Rivenson et al., 2019b; Kandel et al., 2020; Bai et al., 2022; Bai et al., 2023).

We believe the rapid expansion of such virtual staining applications and their integration with other microscopy-enhancing network models will invariably lead to transformative opportunities in biomedical imaging.

Virtual staining

Several virtual staining models have already been successfully designed and deployed, encompassing a variety of organ and staining types (Rivenson et al., 2019a; Rivenson et al., 2019b; Rivenson et al., 2020; Zhang et al., 2020; Li et al., 2021; Bai et al., 2022; Pillar and Ozcan, 2022). It has been shown that tissue biopsy images obtained with holographic microscopy or autofluorescence can be used to virtually generate the equivalents of standard histochemical stains using deep learning algorithms. In many cases, the networks involve a supervised form of the conditional generative adversarial network “GAN” (Goodfellow et al., 2020) (Figures 1A,B), which consists of a generator and a discriminator competing in a zero-sum setting. Such virtually stained slides have shown very good fidelity with their histologically stained counterparts when evaluated by pathologists (Pillar and Ozcan, 2022). This virtual staining technique greatly reduces manual labor and the costs associated with customary laboratory preparations of chemically stained tissue.

FIGURE 1.

FIGURE 1

(A) Example of virtual staining of a QPI image to digitally generate an H&E brightfield image (Rivenson et al., 2019a), scale bar 50 μm, (B) example of virtual staining of an autofluorescence image to generate a HER2 brightfield image (Bai et al., 2022), scale bar 100 µm, (C) example of inferring special stains from an existing H&E stain (de Haan et al., 2021), scale bar 50 μm, (D) example of virtual staining of a blood smear quantitative phase image to digitally generate Wright’s stain (Fanous et al., 2022), scale bar 25 μm, (E) example of virtual staining of QPI cell images with a DAPI nuclear stain (Kandel et al., 2020), scale bar 25 µm.

Recent advances in this emerging field also include virtual staining of label-free images obtained in vivo (Li et al., 2021) and stain-to-stain transformations, e.g., generating Masson’s trichome stain from the image of hematoxylin and eosin (H&E) (de Haan et al., 2021) stained tissue, as shown in Figure 1C, and even blending various stains into an intelligent amalgam for diagnostic optimization (Zhang et al., 2020).

Virtual staining has also been used on label-free images of blood smears to artificially generate the Giemsa (Kaza et al., 2022) or Wright’s stain (Fanous et al., 2022) (Figure 1D), which are commonly used to diagnose leukocyte and erythrocyte disorders. Extension of virtual staining to various fluorescent probes has also been developed to specifically detect subcellular structures of interest without the need for fluorescent tags (Ounkomol et al., 2018; Kandel et al., 2020; Jo et al., 2021; He et al., 2022; Hu et al., 2022), with deep neural networks involving mostly U-Net architectures (Kandel et al., 2020). In one such experiment, the growth of the nucleus and cytoplasm of SW480 cells was assessed over many days by applying the computed fluorescence maps back to the corresponding QPI data (Kandel et al., 2020) (Figure 1E).

Another study used virtual staining to generate semantic segmentation maps from computationally inferred fluorescence images in live, unlabeled brain cells that were subsequently utilized to decipher cellular compartments (Kandel et al., 2021). The time-lapse development of hippocampal neurons was further studied using these synthetic fluorescence signals, emphasizing the connections between cellular dry mass generation and the movements of biomolecules inside the nucleus and neurites. This technique allowed for continuous recordings of live samples without deleterious fluorescent elements.

Table 1 provides an overview of some of these virtual staining approaches, including the tested sample types and the specific advantages they offer in addition to the cost, labor, and time savings compared with traditional chemical staining methods.

TABLE 1.

Different virtual staining methods with their applications, tested samples and specific added benefits.

Virtual staining application Virtual staining type Tested samples Benefits*
Clinical use Label-free to colorimetric histology stains Salivary gland, thyroid, liver, lung, kidney Rivenson (2018), carotid Li et al. (2020); Zhang et al. (2022a), ovarian Meng et al. (2021) and skin tissue Rivenson et al. (2019a) Uniform and repeatable staining, removal of human-induced staining artifacts, permits digital stain multiplexing on the same tissue section
Label-free to immunohistochemical (IHC) stains Breast Bai et al. (2022) and gastric tissue Hong et al. (2021) Label-free biomarker for diagnostics/prognostics, tumor-stroma measurements
Label-free in vivo virtual staining Skin tissue Li et al. (2021) Noninvasive, biopsy-free staining of skin tissue
Label-free to cytology stains Blood smears Fanous et al. (2022); Kaza et al. (2022) and sperm cells Nygate et al. (2020) Simpler workflow, quantitative cell properties, less toxicity
Enhanced diagnosis Stain-to-stain transformations Kidney tissue de Haan et al. (2021) Additional contrast to tissue components, highlights cells absent in deeper sections
Digital stain blending Kidney tissue Zhang et al. (2020) Optimized diagnosis, digital creation of new types of stains
Research Label-free to fluorescent stains Colorectal cancer cells Kandel et al. (2020), HeLa Jo et al. (2021) and CHO cells He et al. (2022); Hu et al. (2022), embryonic kidney cells Ounkomol et al. (2018), and viruses Goswami et al. (2021) Reduced photodamage, live measurements, high-throughput

*All virtual staining methods reduce cost, waste, labor, and assay time.

Discussion

Over the past century, light microscopy has undergone a remarkable and profound transformation. It has transitioned from being predominantly descriptive and qualitative to becoming a potent tool capable of uncovering novel phenomena and elucidating intricate molecular mechanisms through a synergistic visual and quantitative approach. One key driving factor behind these advancements has been the development of numerous immunohistochemical (IHC) stains that effectively highlight specific epitopes within cells. These IHC stains have significantly enhanced diagnostic capabilities in research and clinical pathology. However, in challenging cases, several IHC stains are often employed, necessitating the use of multiple tissue slides for analysis. This becomes a bottleneck as tissue biopsies are becoming smaller in size, and there is a growing need to harness new technologies that can extract more information from limited tissue samples. With its non-destructive nature, alternative label-free optical modalities, when combined with virtual staining, hold the potential to revolutionize the histology field by enabling multiple stains from a single tissue section. This advancement opens doors for more accurate diagnosis, even when working with relatively small tissue fragments. Furthermore, a notable decrease in required reagents and chemicals, including multiple specific antibodies, can prove highly advantageous for small laboratories that lack the financial means to maintain an ever-expanding inventory of diagnostic antibodies.

The overall processing time for a typical IHC stain typically spans a couple of days. Nevertheless, certain clinical situations such as transplanted organs with suspected rejection or rapidly growing tumors necessitate a significantly expedited pathological report. As for the virtual staining of whole slide images (WSI), the latest cutting-edge techniques can accomplish this process within minutes. Customizing the image acquisition system and digital processing hardware could further accelerate this operation and simplify the whole measurement process. For instance, it has been shown that GAN networks can be constructed and trained to deblur out-of-focus images with high reliability in frames with axial offsets of up to +/− 5 µm from the image plane (Luo et al., 2021). To accelerate the tissue imaging process, which often consists of frequent focus adjustments during the scanning of a WSI, cascaded networks have been assembled to first restore the sharpness of defocused images that randomly appear during the slide scanning process, and then digitally perform virtual staining on these autofocused images (Zhang Y. et al., 2022). This two-step tactic enabled by a cascade of autofocusing and virtual staining neural networks is an example of how deep learning can be used to enhance not only the sample preparation and staining processes, but also the measurement, i.e., the image acquisition step.

Similarly, digital staining could potentially be coupled with the recently devised motion-blur reconstruction method named GANscan (Fanous and Popescu, 2022; Rivenson and Ozcan, 2022). This technique scans tissue slides in a continuous manner at 30-times the speed of traditional microscopy scanning, and subsequently corrects for the speed-induced motion-blur effect through a GAN-trained network. If the inputs are images generated by a label-free contrast mechanism such as QPI or autofluorescence, the results of the model could thereafter be digitally stained. This, again, could constitute a cascaded neural network architecture, first handling the deblurring operation due to rapid scanning of the tissue sample, and then virtual staining of the deblurred samples from label-free endogenous contrast to a desired virtual stain.

Another deep learning operation that can be advantageously paired with virtual staining is the enhancement of spatial resolution. It has been shown that deep learning models can be trained to convert diffraction-limited confocal microscopy images into super-resolved stimulated emission depletion (STED) microscopy equivalent images (Wang et al., 2019). To our knowledge, a concept that has not yet been realized is achieving super-resolved quantitative phase imaging through the supervised learning of fluorescent-to-phase modalities, flipping the typical direction of transformation using labeled samples. Coupling such a virtual super-resolution network with digital staining could, in principle, allow one to obtain super-resolution brightfield H&E images from ordinary label-free QPI acquisitions.

Overall, virtually transforming one imaging modality into another, along with advances in deep learning tools, has been the boon of many meaningful microscopy innovations in recent years. And there are multiple circumstances in which such a strategy of cross-modality image transformations is still unexplored or may benefit from further research.

Models may be designed to fix the various imperfections of a sample, whether optical or physical, and could thereafter be virtually stained. A consecutive GAN network would first handle artifact reconstructions/corrections, and then the stain of choice would be digitally rendered. It is also worth noting that implementing a system that enables rapid and consistent imaging, correction, and virtual staining of tissue samples would significantly enhance stain uniformity/repeatability. This is particularly crucial considering the lab-based biases present in extensive and reputable databases, such as the digital image collection of The Cancer Genome Atlas (TCGA) (Dehkharghanian et al., 2023).

The cardinal challenges to such strategies are twofold: first, a copious amount of data is required for acceptable results. Enough instances need to be included to handle the various anomalies and differences of each case; second, as this is primarily a supervised learning approach, the image pairs need to be very well registered, which might be tedious and require manual inspection and quality assurance during the training data preparation (which is a one-time effort).

Conclusion

Virtual staining has demonstrated powerful capabilities using various modes of microscopy and will likely be implemented more and more in different bioimaging scenarios, steadily modernizing the industry altogether. The ability of virtual staining to accurately highlight tissue morphology while conserving tissue, reducing costs, and expediting turnaround time has the potential to revolutionize traditional histopathology workflows. However, for a truly disruptive virtual staining-based digitization of the well-established branches and subspecialties of pathology to occur, the technologies spanning both ends of the histological process (from sample acquisition to physician examination) need to be not only highly ergonomic, comprehensive and consistent, but also affordable and compatible with different forms of microscopy and slide scanner devices that are commercially available.

Acknowledgments

AO Lab acknowledges the support of NIH P41, The National Center for Interventional Biophotonic Technologies.

Data availability statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Conflict of interest

AO is the co-founder of a company (Pictor Labs) that commercializes virtual staining technologies.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Alturkistani H. A., Tashkandi F. M., Mohammedsaleh Z. M. (2016). Histological stains: A literature review and case study. Glob. J. health Sci. 8, 72. 10.5539/gjhs.v8n3p72 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bai B., Wang H., Li Y., de Haan K., Colonnese F., Wan Y., et al. (2022). Label-free virtual HER2 immunohistochemical staining of breast tissue using deep learning. BME Front. 2022. 10.34133/2022/9786242 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Bai B., Yang X., Li Y., Zhang Y., Pillar N., Ozcan A. (2023). Deep learning-enabled virtual histological staining of biological samples. Light Sci. Appl. 12, 57. 10.1038/s41377-023-01104-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Burch C., Stock (1942). Phase-contrast microscopy. J. Sci. Instrum. 19, 71–75. 10.1088/0950-7671/19/5/302 [DOI] [Google Scholar]
  5. de Haan K., Zhang Y., Zuckerman J. E., Liu T., Sisk A. E., Diaz M. F. P., et al. (2021). Deep learning-based transformation of H&E stained tissues into special stains. Nat. Commun. 12, 4884–4913. 10.1038/s41467-021-25221-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Dehkharghanian T., Bidgoli A. A., Riasatian A., Mazaheri P., Campbell C. J. V., Pantanowitz L., et al. (2023). Biased data, biased AI: Deep networks predict the acquisition site of TCGA images. Diagn. Pathol. 18, 67–12. 10.1186/s13000-023-01355-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Fanous M. J., He S., Sengupta S., Tangella K., Sobh N., Anastasio M. A., et al. (2022). White blood cell detection, classification and analysis using phase imaging with computational specificity (PICS). Sci. Rep. 12, 20043–20110. 10.1038/s41598-022-21250-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Fanous M. J., Popescu G. (2022). GANscan: Continuous scanning microscopy using deep learning deblurring. Light Sci. Appl. 11, 265. 10.1038/s41377-022-00952-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Goodfellow I., Pouget-Abadie J., Mirza M., Xu B., Warde-Farley D., Ozair S., et al. (2020). Generative adversarial networks. Commun. ACM 63, 139–144. 10.1145/3422622 [DOI] [Google Scholar]
  10. Goswami N., He Y. R., Deng Y. H., Oh C., Sobh N., Valera E., et al. (2021). Label-free SARS-CoV-2 detection and classification using phase imaging with computational specificity. Light Sci. Appl. 10, 176. 10.1038/s41377-021-00620-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. He Y. R., He S., Kandel M. E., Lee Y. J., Hu C., Sobh N., et al. (2022). Cell cycle stage classification using phase imaging with computational specificity. ACS photonics 9, 1264–1273. 10.1021/acsphotonics.1c01779 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Hong Y., Heo Y. J., Kim B., Lee D., Ahn S., Ha S. Y., et al. (2021). Deep learning-based virtual cytokeratin staining of gastric carcinomas to measure tumor–stroma ratio. Sci. Rep. 11, 19255. 10.1038/s41598-021-98857-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Hu C., He S., Lee Y. J., He Y., Kong E. M., Li H., et al. (2022). Live-dead assay on unlabeled cells using phase imaging with computational specificity. Nat. Commun. 13, 713. 10.1038/s41467-022-28214-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Jo Y., Cho H., Park W. S., Kim G., Ryu D., Kim Y. S., et al. (2021). Label-free multiplexed microtomography of endogenous subcellular dynamics using generalizable deep learning. Nat. Cell Biol. 23, 1329–1337. 10.1038/s41556-021-00802-x [DOI] [PubMed] [Google Scholar]
  15. Kandel M. E., He Y. R., Lee Y. J., Chen T. H. Y., Sullivan K. M., Aydin O., et al. (2020). Phase Imaging with Computational Specificity (PICS) for measuring dry mass changes in sub-cellular compartments. Nat. Commun. 11, 6256–6310. 10.1038/s41467-020-20062-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Kandel M. E., Kim E., Lee Y. J., Tracy G., Chung H. J., Popescu G. (2021). Multiscale assay of unlabeled neurite dynamics using phase imaging with computational specificity. ACS sensors 6, 1864–1874. 10.1021/acssensors.1c00100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Kaza N., Ojaghi A., Robles F. E. (2022). Virtual staining, segmentation, and classification of blood smears for label-free hematology analysis. BME Front. 2022. 10.34133/2022/9853606 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Lang W. (1982). Nomarski differential interference-contrast microscopy. Carl Zeiss Oberkochen. [Google Scholar]
  19. Li D., Hui H., Zhang Y., Tong W., Tian F., Yang X., et al. (2020). Deep learning for virtual histological staining of bright-field microscopic images of unlabeled carotid artery tissue. Mol. imaging Biol. 22, 1301–1309. 10.1007/s11307-020-01508-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Li J., Garfinkel J., Zhang X., Wu D., Zhang Y., de Haan K., et al. (2021). Biopsy-free in vivo virtual histology of skin using deep learning. Light Sci. Appl. 10, 233–322. 10.1038/s41377-021-00674-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Lichtman J. W., Conchello J.-A. (2005). Fluorescence microscopy. Nat. methods 2, 910–919. 10.1038/nmeth817 [DOI] [PubMed] [Google Scholar]
  22. Luo Y., Huang L., Rivenson Y., Ozcan A. (2021). Single-shot autofocusing of microscopy images using deep learning. ACS Photonics 8, 625–638. 10.1021/acsphotonics.0c01774 [DOI] [Google Scholar]
  23. Majeed H., Keikhosravi A., Kandel M. E., Nguyen T. H., Liu Y., Kajdacsy-Balla A., et al. (2019). Quantitative histopathology of stained tissues using color spatial light interference microscopy (cSLIM). Sci. Rep. 9, 14679–14714. 10.1038/s41598-019-50143-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Majeed H., Sridharan S., Mir M., Ma L., Min E., Jung W., et al. (2017). Quantitative phase imaging for medical diagnosis. J. Biophot. 10, 177–205. 10.1002/jbio.201600113 [DOI] [PubMed] [Google Scholar]
  25. Meng X., Li X., Wang X. (2021). A computationally virtual histological staining method to ovarian cancer tissue by deep generative adversarial networks. Comput. Math. Methods Med. 2021, 1–12. 10.1155/2021/4244157 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Monici M. (2005). Cell and tissue autofluorescence research and diagnostic applications. Biotechnol. Annu. Rev. 11, 227–256. 10.1016/S1387-2656(05)11007-2 [DOI] [PubMed] [Google Scholar]
  27. Nygate Y. N., Levi M., Mirsky S. K., Turko N. A., Rubin M., Barnea I., et al. (2020). Holographic virtual staining of individual biological cells. Proc. Natl. Acad. Sci. 117, 9223–9231. 10.1073/pnas.1919569117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Ounkomol C., Seshamani S., Maleckar M. M., Collman F., Johnson G. R. (2018). Label-free prediction of three-dimensional fluorescence images from transmitted-light microscopy. Nat. methods 15, 917–920. 10.1038/s41592-018-0111-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Park Y., Depeursinge C., Popescu G. (2018). Quantitative phase imaging in biomedicine. Nat. Photonics 12, 578–589. 10.1038/s41566-018-0253-x [DOI] [Google Scholar]
  30. Pillar N., Ozcan A. (2022). Virtual tissue staining in pathology using machine learning. Expert Rev. Mol. Diagnostics 22, 987–989. 10.1080/14737159.2022.2153040 [DOI] [PubMed] [Google Scholar]
  31. Rivenson Y., de Haan K., Wallace W. D., Ozcan A. (2020). Emerging advances to transform histopathology using virtual staining. BME Front. 2020. 10.34133/2020/9647163 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Rivenson Y. (2018). Deep learning-based virtual histology staining using auto-fluorescence of label-free tissue. arXiv preprint arXiv:1803.11293 . [Google Scholar]
  33. Rivenson Y., Liu T., Wei Z., Zhang Y., de Haan K., Ozcan A. (2019a). PhaseStain: The digital staining of label-free quantitative phase microscopy images using deep learning. Light Sci. Appl. 8, 23–11. 10.1038/s41377-019-0129-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Rivenson Y., Ozcan A. (2022). Deep learning accelerates whole slide imaging for next-generation digital pathology applications. Light Sci. Appl. 11, 300–303. 10.1038/s41377-022-00999-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Rivenson Y., Wang H., Wei Z., de Haan K., Zhang Y., Wu Y., et al. (2019b). Virtual histological staining of unlabelled tissue-autofluorescence images via deep learning. Nat. Biomed. Eng. 3, 466–477. 10.1038/s41551-019-0362-y [DOI] [PubMed] [Google Scholar]
  36. Wang H., Rivenson Y., Jin Y., Wei Z., Gao R., Günaydın H., et al. (2019). Deep learning enables cross-modality super-resolution in fluorescence microscopy. Nat. methods 16, 103–110. 10.1038/s41592-018-0239-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Zhang G., Ning B., Hui H., Yu T., Yang X., Zhang H., et al. (2022a). Image-to-images translation for multiple virtual histological staining of unlabeled human carotid atherosclerotic tissue. Mol. Imaging Biol. 24, 31–41. 10.1007/s11307-021-01641-w [DOI] [PubMed] [Google Scholar]
  38. Zhang Y., de Haan K., Rivenson Y., Li J., Delis A., Ozcan A. (2020). Digital synthesis of histological stains using micro-structured and multiplexed virtual staining of label-free tissue. Light Sci. Appl. 9, 78–13. 10.1038/s41377-020-0315-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Zhang Y., Huang L., Liu T., Cheng K., de Haan K., Li Y., et al. (2022b). Virtual staining of defocused autofluorescence images of unlabeled tissue using deep neural networks. Intell. Comput. 2022. 10.34133/2022/9818965 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.


Articles from Frontiers in Bioinformatics are provided here courtesy of Frontiers Media SA

RESOURCES