Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2020 May 14;10:7968. doi: 10.1038/s41598-020-64937-x

Artificial Intelligence in Multiphoton Tomography: Atopic Dermatitis Diagnosis

Pedro Guimarães 1,, Ana Batista 2, Michael Zieger 3, Martin Kaatz 3, Karsten Koenig 2,4
PMCID: PMC7224284  PMID: 32409755

Abstract

The diagnostic possibilities of multiphoton tomography (MPT) in dermatology have already been demonstrated. Nevertheless, the analysis of MPT data is still time-consuming and operator dependent. We propose a fully automatic approach based on convolutional neural networks (CNNs) to fully realize the potential of MPT. In total, 3,663 MPT images combining both morphological and metabolic information were acquired from atopic dermatitis (AD) patients and healthy volunteers. These were used to train and tune CNNs to detect the presence of living cells, and if so, to diagnose AD, independently of imaged layer or position. The proposed algorithm correctly diagnosed AD in 97.0 ± 0.2% of all images presenting living cells. The diagnosis was obtained with a sensitivity of 0.966 ± 0.003, specificity of 0.977 ± 0.003 and F-score of 0.964 ± 0.002. Relevance propagation by deep Taylor decomposition was used to enhance the algorithm’s interpretability. Obtained heatmaps show what aspects of the images are important for a given classification. We showed that MPT imaging can be combined with artificial intelligence to successfully diagnose AD. The proposed approach serves as a framework for the automatic diagnosis of skin disorders using MPT.

Subject terms: Machine learning, Optical imaging, Diagnosis, Medical imaging, Skin diseases

Introduction

The human skin is affected by many prevalent diseases, typically widespread across the population and with an immense impact on the patients’ day-to-day life. However, the similarity between lesions hinders disease diagnostics.

Atopic Dermatitis (AD) is the most common chronic inflammatory disease, with prevalence up to 20% in developed countries. It often develops during early childhood, and is recurred throughout the patient’s lifespan1, highly influencing their and their families quality of life2. AD is characterized by eczematous lesions. However, clinical features manifested are considerably heterogenous, varying with respect to lesion distribution, morphology, intensity, and duration1,3. This heterogeneity often hinders the clinician ability to correctly diagnose it.

Multiphoton tomography (MPT) allows non-invasive, label-free, sub-cellular, and 3D-resolved imaging of the human skin. It can simultaneously provide information on tissue morphology and metabolism with high lateral and axial resolutions. Morphology contrast is achieved by the endogenous fluorophores autofluorescence (AF) and collagen second-harmonic generation (SHG). Metabolic information is obtained using nicotinamide adenine dinucleotide and nicotinamide adenine dinucleotide phosphate (NAD(P)H) AF lifetime. During energy production, NAD(P)H undergoes oxidation/reduction mediated by proteins4,5. Based on its AF lifetime, free and protein-bound components of NAD(P)H can be separated5. The ratio between these components is an indirect measure of the cell’s metabolic activity4,6.

The potential of MPT for skin diagnosis is clear and it has already been demonstrated for diseases such as basal and squamous cell carcinomas710, skin melanoma1012 and even AD1317. Cell irregularities, subcellular changes, like perinuclear accumulation of mitochondria, and metabolic changes are some of the findings observed in the inflamed skin of AD patients14,15. Nonetheless, its potential has yet to be fully realized.

In recent years, artificial intelligence-based approaches for image analysis have been exponentially researched and developed. Convolutional neural networks (CNNs) form a subset of deep learning algorithms that take advantage of the relationships between neighboring pixels combining them in consecutive layers to represent exponentially more complex patterns. CNNs are the de-facto networks for image analysis, and have shown strong diagnostic performance when applied to medical images from different fields such as cardiology18,19, neurology20,21, gastroenterology22,23, and dermatology3,24,25.

In this study, we demonstrate the feasibility of artificial intelligence for the diagnosis of AD from MPT images. We introduce an accurate and reliable deep learning algorithm for the identification of images with living cells and subsequent diagnosis, removing any operator dependency. We showed that these approaches may be used to fully explore the potential of MPT. This initial work serves as framework for other skin diseases.

Material and Methods

Data Source

Data collection was conducted by JenLab GmbH in collaboration with the SRH Wald-Klinikum Gera. This study was approved by the Ethics Committee of the Medical Association of Thuringia, Germany (EUDAMED No: CIV- 17-12-022506) and was conducted following the principles of the World Medical Association Declaration of Helsinki. Informed consent was obtained from all participants.

The analyzed image dataset contains 3,663 patches pulled from a total of 21 image z-stacks acquired from 10 different anonymized subjects. Up to three non-overlapping volumes of the forearm were imaged per subject (two stacks average). In total, six subjects were diagnosed with AD and four were healthy volunteers. A detailed distribution of all the patches according to class is shown in Table 1.

Table 1.

Detailed distribution of image patches according to class.

Healthy Atopic Dermatitis Total
With Living Cells 566 853 1,419
Without Living Cells 1,018 1,226 2,244
Total 1,584 2,079 3,663

Each z-stack consists of sequential 160 × 160 µm2 enface images of the skin obtained with 5 µm step in between them, up to a depth of 100 µm. Image resolution was about 300 nm laterally and 1–2 µm axially. After initial fluorescence lifetime analysis, each image was cropped to nine 200 × 200 pixels patches (equivalent to an area of 62.5 × 62.5 µm2), with three concatenated channels, consisting of AF intensity, mean AF lifetime, and the ratio between the short (a1) and long (a2) relative contributions of AF lifetime components. Examples of the three input channels are shown in Fig. 1. Each patch was also evaluated by an independent expert to detect the presence of living cells. Subjects of both sexes were included, with ages varying between 19 and 78 years old for AD patients and between 27 and 53 for healthy volunteers.

Figure 1.

Figure 1

Examples of the three image types to be concatenated for input to the deep learning model. Left – autofluorescence (AF) intensity image; Middle – mean AF lifetime; Right – ratio between the short (a1) and long (a2) relative contributions of AF lifetime components. FLIM images were analyzed using the commercial software SPCImage v8.0 (Becker & Hickl GmbH, Berlin, Germany).

In vivo MPT Imaging Acquisition of the Human Skin

In vivo image acquisition was performed using the multiphoton tomograph MPTflex-CARS (JenLab GmbH, Berlin, Germany). This device has been described in detail elsewhere26. Briefly, this is a movable and flexible MPT imaging device composed by an optoelectronic housing that encompasses the laser and other optoelectronic components, an articulated mirror-arm to guide the laser light, and a flexible 360° scan-detector head. Endogenous fluorophores were excited using an 80 MHz tunable near-infrared (NIR) Ti:Sapphire 100 femtosecond (fs) oscillator, tuned at 800 nm. A high numerical aperture (NA) objective (40×1.3 NA) focused the laser light onto the sample and collected generated signals in reflection geometry. Galvanometric x-y scanners and a z-stepper motor enable x-y-z imaging. A photomultiplier tube (PMT) coupled with a time-correlated single photon counting (TCSPC) module for fluorescence lifetime imaging (FLIM) detected fluorescence. Simultaneously, a second PMT detected second-harmonic generation (SHG) signals from collagen. With this device additional acquisition of coherent anti-stokes Raman scattering (CARS) signals is possible26.

Fluorescence Lifetime Analysis

FLIM images were analyzed using the commercial software SPCImage v8.0 (Becker & Hickl GmbH, Berlin, Germany). AF lifetime data was retrieved by fitting it histogram to a bi-exponential function in the form:

F(t)=a1etτ1+a2etτ2, 1

where F(t) is the AF intensity at time t, and a1 and a2 are the relative contribution of the AF lifetime components τ1 and τ2, respectively. Pixel binning was applied to guarantee the necessary number of photon-counts for accurate fitting. After the selection of the fitting parameters, all FLIM images of the skin were analyzed using batch processing, without user intervention.

AD-induced changes were evaluated depth-wise based on the ratio between a1 and a2 (i.e., a1/a2) and the tissue mean AF lifetime (τm), obtained as:

τm=a1τ1+a2τ2, 2

where a1+a2=1. The analysis of the living cell layers was performed after the depth-wise alignment of the stratum granulosum of all z-stacks. The average mean AF lifetime and a1/a2 ratio were computed and recorded at each depth for AD and control subjects. Being composed by dead cells (corneocytes), the stratum corneum was disregarded in this analysis. Statistical differences between healthy and AD-affected skin were assessed using the non-parametric test Mann-Whitney U test (p < 0.5).

Pixel resolved images of mean AF lifetime and a1/a2 ratio were exported and concatenated with the AF intensity.

Atopic Dermatitis Automatic Diagnosis

Although the diagnostic possibilities of MPT imaging have been shown716, currently the analysis of these images is still largely operator dependent. Factors such as different regions-of-interest and analysis parameters, lead to substantially different results. Here, we propose a fully automatic protocol capable of showing the full potential of these methods. For this reason, deep learning as an end-to-end approach is the perfect fit.

Deep learning Models

Image z-stacks are captured from different depths, imaging several layers and covering areas with and without living cells. Previous works have reported changes in the metabolism and morphology of living cells in AD13,14. Our approach starts by identifying whether these cells are present in a given image, and if so, evaluates if the imaged skin is diseased. As so, two different CNN models have been developed and trained.

Deep learning models work in contrast with traditional machine-learning approaches: features are not engineered and fed for classification, instead, they are computed incrementally by the model itself. It can create, layer to layer, progressively more complex abstract representations (i.e., features), having no theoretical restraint to learn any representation. Nonetheless, this requires a large amount of training data. Deep learning models are prone to overfitting, especially with limited training data, i.e., the model learns from patterns that are specific to the training and do not generalize to other data. To overcome this, three actions were taken: (i) we used artificial data augmentation by image rotation, horizontal and vertical image reflection, and scaling, (ii) dropout regularization added to the last fully connected layer of the models (the removal of random nodes approximates the behavior of an ensemble of multiple network architectures that has been shown to reduce overfitting27), and (iii) we fine-tuned pre-trained CNNs, i.e., transfer-learning. Pre-trained weights are used to initialize the network, thus improving the stability and performance of the model. In this work, we used DenseNet architecture28, with weights pre-trained on ImageNet29, a dataset with over 14 million labelled images belonging to about 22,000 classes. We adapted the model by replacing the last fully connected and output layer.

Model Training and Testing

The performance of the proposed algorithm was evaluated using subject-wise leave-one-out cross-validation. We performed a total of 10 cross-validation runs dividing the data into train, tune and test sets. To ensure complete independency between all the sets, for each run the composition of each set was changed, i.e. a new model was created. All images from one of the subjects were left out for testing, the images from another were left out for tuning, while the remainder images were used for training. Each subject was selected only once for testing and could not ever be simultaneously in the testing and training sets. Thus, there is no selection bias by assigning specific testing sets.

Only the training set is artificially augmented. The models were fine-tuned using stochastic gradient descent with a small learning rate and binary cross-entropy was defined as the objective function. Class dependent weighting of the objective function was applied in training to account for class unbalance. The test set is classified using the best performing hyperparameter combination, as assessed by grid-search with early-stopping, evaluated in the tuning set. The evaluated hyperparameters include dropout rate, learning rate, and momentum. All the tested software was implemented in Python v3.5. Models were implemented with Tensorflow v2.030.

Variational Dropout for Uncertainty Estimation

For the application of neural networks in the medical field it is imperative to quantify uncertainty, and model output and confidence are not directly correlated31. In this study, variational dropout was applied to estimate uncertainty. The idea is to run dropout during testing and perform multiple prediction calls31,32. As aforementioned, by turning off several units of the network at random, the multiple calls approximate the behavior of an ensemble of multiple models27. The results can then be interpreted as a Bayesian probability distribution31,32. In our implementation we performed 50 prediction calls averaging the probabilities obtained. Uncertainty was obtained as the standard deviation of those probabilities. Bayesian probability and variational dropout gives us a mathematically sound approach for uncertainty estimation without forfeiting computational cost or prediction accuracy31.

Deep Taylor Decomposition

One of the main drawbacks of the application and acceptance of non-linear methods such as deep learning and some other machine-learning approaches in the medical field, is the ‘black box’ nature of these methods.

Several approaches exist to improve the interpretability of deep learning methods33. In this work, deep Taylor decomposition was applied34. The rational is to decompose the final decision into individual contributions of each given input by means of relevance backpropagation, i.e., trace-back contributions between layers from output to input. The result are heatmaps representing the relative relevance of each pixel to the final classification.

Classification Performance Metrics

To evaluate our models’ performance, accuracy, balanced accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and F-score were calculated. Receiving operating characteristics (ROC) and precision-recall curves were evaluated. The area under the curve (AUC) was also assessed for both curves. Metrics were computed for the detection of living cells (in the entire dataset) and subsequent disease detection (for images with a positive classification for the presence of living cells). Confidence intervals for all the obtained metrics were also computed from the multiple model calls.

Results

Preliminary Analysis: Autofluorescence Intensity and Lifetime

Morphologic and metabolic AD-induced changes can be detected by MPT imaging. Figure 2a,b show tree-dimensional (3D) representations of the AF intensity (green) and SHG (red) of healthy and AD-affected skin, respectively. Highlighted regions show cell morphology. Corresponding FLIM images are shown in Fig. 2c,d. In the outermost living cell layers of the skin epidermis, AF signals are generated mainly by NAD(P)H located in the cells’ mitochondria. Therefore, changes in mitochondria arrangement within the stratum granulosum cells caused by AD, can be observed. In healthy skin, mitochondria appear to be evenly distributed in the cell cytoplasm (Fig. 2a). In contrast, in AD-affected skin, mitochondria appear to accumulate perinuclearly (Fig. 2b).

Figure 2.

Figure 2

Autofluorescence intensity and lifetime analysis. Three-dimensional representation of the skin autofluorescence (green) and second-harmonic generation (red) for healthy volunteers (a) and atopic dermatitis patients (b) with highlighted regions showing cell morphology and corresponding FLIM images (c and d, respectively). FLIM images are color-coded for the mean autofluorescence (τm) lifetime as indicated in the color bar and autofluorescence intensity weighted. Scale bar = 20 µm. Changes with depth of τm lifetime (c) and a1/a2 ratio (d) in the living cell layers of healthy and atopic dermatitis diagnosed skin. FLIM images were analyzed using the commercial software SPCImage v8.0 (Becker & Hickl GmbH, Berlin, Germany).

Figure 2e,f show the mean AF lifetimes and the a1/a2 ratio for different depths, respectively. A tendency for lower mean AF lifetimes was observed in the stratum granulosum (depth = 0 µm) of AD-affected skin when compared with healthy skin (Fig. 2e), however statistical significance was not reached. No changes were observed at stratum spinosum (depth = 15 and 25 µm). In the stratum basale, the detection of melanin AF leads to a decrease in the mean AF lifetime (Fig. 2e) and an increase in the a1/a2 ratio (Fig. 2f). This can be observed at approximately 35 µm depth in healthy skin. Due to the increase in epidermal thickness, in AD-affected skin, the stratum basale was not yet reached at this depth and such changes were not observed.

Atopic dermatitis automatic diagnosis

From the total 3,663 patches, averaging 184 per z-stack, 1,419 were considered by a human expert to contain living cells (38.7%, averaging ~71 per z-stack), covering the stratum granulosum, stratum spinosum, and stratum basale layers. Images presented varying levels of quality. Nevertheless, all images were considered.

The deep learning model used to detect living cells performed well. ROC and precision-recall curves were computed and are shown in Fig. 3a,b, respectively. The AUC values achieved were 0.968 ± 6.0 × 10−4, and 0.955 ± 7.3 × 10−4, for the ROC and precision-recall curves, respectively.

Figure 3.

Figure 3

Performance curves. Receiving operating characteristics (a) and precision-recall (b) curves for the living cells and atopic dermatitis detection tasks. The range for each curve is represented in gray. The area under each curve (AUC) ± standard deviation values are showed in the graphic’s legend.

In total, 1,419 images were classified as containing living cells. These were then passed to the second network (AD classification). From those, 840 of them were acquired from subjects diagnosed with AD (59.5%). The ROC and precision-recall curves for AD classification show the excellent performance achieved (Fig. 3). AUC values of 0.994 ± 3.8 × 10−4 and 0.996 ± 2.2 × 10−4 were computed, respectively.

Table 2 summarizes the remaining performance metrics for both classification tasks. Cell detection achieved an accuracy of 0.910 ± 0.002 with a higher specificity value when compared to its sensitivity (0.927 ± 0.002 vs 0.884 ± 0.004). False positives were passed down to AD classification (about 11.6% of all positives). Despite this, our approach accurately diagnosed correctly ~97% of all images, achieving a balance between sensitivity and specificity (0.966 ± 0.003 and 0.977 ± 0.003, respectively). Inter-subject and inter-stack variation were assessed by the accuracy standard deviation (Table 3). Results were consistent between z-stacks and subjects as shown by the low values obtained.

Table 2.

Classification Performance. Performance metrics ± standard deviation values for the living cells and atopic dermatitis detection tasks. PPV – Positive predictive value; NPV – Negative predictive value.

Living Cells Atopic Dermatitis
Accuracy 0.910 ± 0.002 0.970 ± 0.002
Balanced Accuracy 0.906 ± 0.003 0.972 ± 0.003
Sensitivity 0.884 ± 0.004 0.966 ± 0.003
Specificity 0.927 ± 0.002 0.977 ± 0.003
PPV 0.884 ± 0.003 0.984 ± 0.002
NPV 0.927 ± 0.002 0.952 ± 0.004
F1-score 0.927 ± 0.001 0.964 ± 0.002

Table 3.

Performance variation. Inter-stack and inter-subject accuracy standard deviation for the living cells and atopic dermatitis detection tasks.

Inter-Stack Inter-Subject
Living Cells 0.036 0.032
Atopic Dermatitis 0.053 0.063

Interpretability

Deep Taylor decomposition was applied to AD detection. Figure 4 shows the resulting heatmaps for selected examples from healthy and AD subjects. Raw AF intensity image is shown on the left column. On the other columns, AF intensity is overlaid by semi-transparent heatmaps showing: in the middle, the relative relevance of each pixel (color-coded according to the represented color-bar), and on the right, the contribution of each input modality (RGB color model: Red - mean AF lifetime; Green - a1/a2 ratio; Blue - AF intensity).

Figure 4.

Figure 4

Deep Learning Interpretability. Deep Taylor decomposition relevance heatmaps for healthy and diseased subjects. Left – raw autofluorescence (AF) intensity image; Middle – AF intensity overlaid by the relative relevance heatmap color-coded according to the respective color bar; Right – AF intensity overlaid by the relative contribution of each input modality color-coded according to the RGB color model (Red-Green-Blue, as: mean AF lifetime, ratio of AF lifetime components relative contributions, and AF intensity). Transparency of overlaid heatmaps according to magnitude.

These heatmaps show what aspects of the input make the model decide on the disease status of a given subject. As shown, cells are key for AD diagnosis. The model seems to focus on one or a few cells in each image. AF intensity and mean AF lifetime (red and blue, respectively) have a high contribution in almost all images, while the a1/a2 ratio contribution is quite dispersed. Regions with high AF intensity in the cell’s cytoplasm seem to be particularly of interest. As above mentioned, these correspond to mitochondria. The contribution of the a1/a2 ratio (displayed in green) is specifically high for some of the non-AD images and almost non-existent in AD images.

Discussion

Using MPT, multiple layers of the skin can be investigated non-invasively with sub-cellular resolution. The nonlinear excitation of endogenous fluorophores enables its analysis at multiple depths without the need for biopsies. The morphology of each layer is assessed using the AF intensity and SHG signals from collagen fibers. In addition, due to the role NAD(P)H in the cell’s metabolic activity, the metabolism can be simultaneously evaluated. It has been previously shown that it can be used to detect inflammation in AD patients1316. Nevertheless, the analysis is still time-consuming and operator dependent. Results are typically specific to a given cell layer (stratum granulosum), and statistical significance is not always achieved13,15,17,35,36.

Morphological and metabolic changes, in accordance to what has been previously reported, were observed in our preliminary analysis. Namely, mitochondria accumulate around the cell’s nucleus in the outermost cell layers of AD patients. This has been previously associated with cell inflammation14,15. A decrease in the mean AF lifetime and an increase in a1/a2 of stratum granulosum cells, without statistical significance, was also observed, indicating decreased metabolic activity16. No changes were observed for the other layers.

In this study, we went a step further and implemented an end-to-end deep learning approach for the automatic classification of AD from MPT data. CNNs have been essential cogs in many of the most recent advances of computer vision and image analysis fields. Given the importance that medical imaging plays in diagnosis, deep learning has the potential to revolutionize many medical areas. In our approach, we combined AF intensity, mean AF lifetime, and a1/a2 ratio, meaning the decision algorithm used both morphological and metabolic information to model AD. The 97.0% classification accuracy achieved by our model from a small imaging area of just 62.5 × 62.5 μm2, shows the potential of combining state-of-the-art imaging and classification methods.

One of the concerns regarding the wide application of machine-learning in the medical field is the lack of information on why, how, and with what certainty a given classification decision is reached. This naturally hinders the possibility of experts verifying the validity of these decisions and detecting/diagnosing a mistake. Deep learning interpretability is of great importance to understand limitations and establish trust in the results.

Uncertainty estimation is of crucial importance. Variational dropout gave us a mathematical foundation to compute model uncertainty for each prediction and allowed us to establish confidence intervals. We also computed relevance heatmaps that show us what aspects of the images are important for a given classification and allow review by experts if necessary. Regions with high AF intensity located within the cytoplasm were prominent. As aforementioned, the main source of AF is NAD(P)H, a coenzyme located mainly in mitochondria. These crucial organelles in the cells’ energy production, are involved in the regulation of inflammation37. Therefore, the high AF intensity pattern is highly relevant for the diagnosis of AD. Metabolic information played also an important part in the model’s decision, as shown by the relative contribution heatmaps. The contribution was much more disperse when compared to AF intensity, coming from inside the cells and regions surrounding them, i.e., less definition. This is to be expected, since binning is required to gather enough photon counts for proper TCSPC curve fitting. The relevance maps indicate that the model is finding similar changes to those described in prior publications1316, integrating and weighting them in an end-to-end manner to reach a decision on the disease status of a given subject.

In summary, we showed that MPT imaging can be combined with artificial intelligence methods to fully explore its potential. We successfully trained a CNN-based approach to detect images with living cells and discriminate AD, achieving a high sensitivity and specificity. By exploring relevance backpropagation to enhance interpretability of the proposed method, we showed that our deep learning model integrates both morphological and metabolic information provided by MPT imaging to derive a highly accurate diagnosis independent of imaging layer or region-of-interest overcoming traditional analysis methods. Moreover, the proposed approach serves not only for AD, but as a framework for the automatic diagnosis of other skin diseases using MPT imaging.

Acknowledgements

The authors would like to acknowledge the support from the European Union’s Horizon 2020 research and innovation program under grant agreement No 726666 (LASER-HISTO) and the support by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) and Saarland University within the funding program Open Access Publishing.

Author contributions

A.B. conducted the fluorescence lifetime analysis. P.G. developed the algorithm. P.G. and A.B. analyzed the data, interpreted the results, and wrote the manuscript. M.Z., M.K. and K.K. provided the data and reviewed the manuscript.

Competing interests

Karsten Koenig is CEO of JenLab GmbH. All other authors declare no conflict of interest.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Weidinger, S., Beck, L. A., Bieber, T., Kabashima, K. & Irvine, A. D. Atopic Dermatitis. in Kanerva’s Occupational Dermatology1, 201–212 (Springer Berlin Heidelberg, 2012).
  • 2.Drucker AM, et al. The Burden of Atopic Dermatitis: Summary of a Report for the National Eczema Association. J. Invest. Dermatol. 2017;137:26–30. doi: 10.1016/j.jid.2016.07.012. [DOI] [PubMed] [Google Scholar]
  • 3.Gustafson, E., Pacheco, J., Wehbe, F., Silverberg, J. & Thompson, W. A Machine Learning Algorithm for Identifying Atopic Dermatitis in Adults from Electronic Health Records. Proc. - 2017 IEEE Int. Conf. Healthc. Informatics, ICHI 2017 83–90, 10.1109/ICHI.2017.31 (2017). [DOI] [PMC free article] [PubMed]
  • 4.Georgakoudi I, Quinn KP. Optical imaging using endogenous contrast to assess metabolic state. Annu Rev Biomed Eng. 2012;14:351–367. doi: 10.1146/annurev-bioeng-071811-150108. [DOI] [PubMed] [Google Scholar]
  • 5.Lakowicz, J. R. Principles of Fluorescence Spectroscopy. (Springer, 2007).
  • 6.Batista A, et al. Assessment of human corneas prior to transplantation using high-resolution two-photon imaging. Investig. Opthalmology Vis. Sci. 2018;59:176–184. doi: 10.1167/iovs.17-22002. [DOI] [PubMed] [Google Scholar]
  • 7.Seidenari S, et al. Diagnosis of BCC by multiphoton laser tomography. Ski. Res. Technol. 2013;19:297–304. doi: 10.1111/j.1600-0846.2012.00643.x. [DOI] [PubMed] [Google Scholar]
  • 8.Balu M, et al. In Vivo Multiphoton Microscopy of Basal Cell Carcinoma. JAMA Dermatology. 2015;151:1068–1075. doi: 10.1001/jamadermatol.2015.0453. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Klemp M, et al. Comparison of morphologic criteria for actinic keratosis and squamous cell carcinoma using in vivo multiphoton tomography. Exp. Dermatol. 2016;25:218–222. doi: 10.1111/exd.12912. [DOI] [PubMed] [Google Scholar]
  • 10.Balu, M. et al. In vivo multiphoton microscopy of human skin. in Multiphoton Microscopy and Fluorescence LifetimeImaging (ed. König, K.) 287–300, 10.1515/9783110429985-017 (De Gruyter, 2018).
  • 11.Balu M, et al. Distinguishing between benign and malignant melanocytic nevi by in vivo multiphoton microscopy. Cancer Res. 2014;70:2688–2697. doi: 10.1158/0008-5472.CAN-13-2582. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Seidenari S, et al. Multiphoton Laser Tomography and Fluorescence Lifetime Imaging of Melanoma: Morphologic Features and Quantitative Data for Sensitive and Specific Non-Invasive Diagnostics. PLoS One. 2013;8(e70682):1–9. doi: 10.1371/journal.pone.0070682. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.König, K. et al. 5D-intravital tomography as a novel tool for non-invasive in-vivo analysis of human skin. in Proc. SPIE7555, 75551I:1–6 (2010).
  • 14.Huck, V. et al. From morphology to biochemical state – intravital multiphoton fluorescence lifetime imaging of inflamed human skin. Sci. Rep 1–12, 10.1038/srep22789 (2016). [DOI] [PMC free article] [PubMed]
  • 15.König K, et al. Translation of two-photon microscopy to the clinic: multimodal multiphoton CARS tomography of in vivo human skin. J. Biomed. Opt. 2020;25:1. doi: 10.1117/1.JBO.25.1.014515. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Mess, C. & Huck, V. Bedside assessment of multiphoton tomography. in Multiphoton Microscopy and Fluorescence LifetimeImaging (ed. König, K.) 425–444, 10.1515/9783110429985-024 (De Gruyter, 2018).
  • 17.König K, Koenig K. Hybrid multiphoton multimodal tomography of in vivo human skin. IntraVital. 2012;1:11–26. doi: 10.4161/intv.21938. [DOI] [Google Scholar]
  • 18.Mortazavi BJ, et al. Analysis of Machine Learning Techniques for Heart Failure Readmissions. Circ. Cardiovasc. Qual. Outcomes. 2016;9:629–640. doi: 10.1161/CIRCOUTCOMES.116.003039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Johnson KW, et al. Artificial Intelligence in Cardiology. Journal of the American College of Cardiology. 2018;71:2668–2679. doi: 10.1016/j.jacc.2018.03.521. [DOI] [PubMed] [Google Scholar]
  • 20.Yang G, et al. DAGAN: Deep De-Aliasing Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. IEEE Trans. Med. Imaging. 2018;37:1310–1321. doi: 10.1109/TMI.2017.2785879. [DOI] [PubMed] [Google Scholar]
  • 21.Pereira S, Pinto A, Alves V, Silva CA. Brain Tumor Segmentation Using Convolutional Neural Networks in MRI Images. IEEE Trans. Med. Imaging. 2016;35:1240–1251. doi: 10.1109/TMI.2016.2538465. [DOI] [PubMed] [Google Scholar]
  • 22.Yamada, M. et al. Development of a real-time endoscopic image diagnosis support system using deep learning technology in colonoscopy. Sci. Rep. 9, (2019). [DOI] [PMC free article] [PubMed]
  • 23.Guimarães P, Keller A, Fehlmann T, Lammert F, Casper M. Deep-learning based detection of gastric precancerous conditions. Gut. 2020;69:4–6. doi: 10.1136/gutjnl-2019-319347. [DOI] [PubMed] [Google Scholar]
  • 24.Zhou, H. et al. Multi-classification of skin diseases for dermoscopy images using deep learning. IST 2017 - IEEE Int. Conf. Imaging Syst. Tech. Proc. 2018-Janua, 1–5 (2018).
  • 25.Shoieb DA, Youssef SM, Aly WM. Computer-Aided Model for Skin Diagnosis Using Deep Learning. J. Image Graph. 2016;4:122–129. doi: 10.18178/joig.4.2.122-129. [DOI] [Google Scholar]
  • 26.Weinigel M, et al. In vivo histology: Optical biopsies with chemical contrast using clinical multiphoton/coherent anti-Stokes Raman scattering tomography. Laser Phys. Lett. 2014;11:055601. doi: 10.1088/1612-2011/11/5/055601. [DOI] [Google Scholar]
  • 27.Srivastava, N., Hinton, G., Krizhevsky, A. & Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research15, (2014).
  • 28.Huang, G. et al. Densely Connected Convolutional Networks. 2017IEEE Conference on Computer Vision andPattern Recognition (CVPR)2017-Janua, 2261–2269 (IEEE).
  • 29.Russakovsky O, et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015;115:211–252. doi: 10.1007/s11263-015-0816-y. [DOI] [Google Scholar]
  • 30.Abadi, M. et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. (2015).
  • 31.Gal, Y. & Ghahramani, Z. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. 33rd Int. Conf. Mach. Learn. ICML 20163, 1651–1660 (2015).
  • 32.Kingma, D. P., Salimans, T. & Welling, M. Variational Dropout and the Local Reparameterization Trick. (2015).
  • 33.Montavon G, Samek W, Müller K-R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 2018;73:1–15. doi: 10.1016/j.dsp.2017.10.011. [DOI] [Google Scholar]
  • 34.Montavon G, Lapuschkin S, Binder A, Samek W, Müller KR. Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recognit. 2017;65:211–222. doi: 10.1016/j.patcog.2016.11.008. [DOI] [Google Scholar]
  • 35.Huck, V. et al. Intravital multiphoton tomography as a novel tool for non-invasive in vivo analysis of human skin affected with atopic dermatitis. in Proc. SPIE7548, 75480B:1–6 (2010).
  • 36.Huck V, et al. Intravital multiphoton tomography as an appropriate tool for non-invasive in vivo analysis of human skin affected with atopic dermatitis. Photonic Ther. Diagnostics VII. 2011;7883:78830R. doi: 10.1117/12.874218. [DOI] [Google Scholar]
  • 37.Meyer, A. et al. Mitochondria: An organelle of bacterial origin controlling inflammation. Frontiers in Immunology9, (2018). [DOI] [PMC free article] [PubMed]

Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES