Abstract
Introduction: Histological grading of cutaneous squamous cell carcinoma (cSCC) is crucial for prognosis and treatment decisions, but manual grading is subjective and time-consuming. Aim: This study aimed to develop and validate a deep learning (DL)-based model for automated cSCC grading, potentially improving diagnostic accuracy (ACC) and efficiency. Materials and Methods: Three deep neural networks (DNNs) with different architectures (AlexNet, GoogLeNet, ResNet-18) were trained using transfer learning on a dataset of 300 histopathological images of cSCC. The models were evaluated on their ACC, sensitivity (SN), specificity (SP), and area under the curve (AUC). Clinical validation was performed on 60 images, comparing the DNNs’ predictions with those of a panel of pathologists. Results: The models achieved high performance metrics (ACC>85%, SN>85%, SP>92%, AUC>97%) demonstrating their potential for objective and efficient cSCC grading. The high agreement between the DNNs and pathologists, as well as among different network architectures, further supports the reliability and ACC of the DL models. The top-performing models are publicly available, facilitating further research and potential clinical implementation. Conclusions: This study highlights the promising role of DL in enhancing cSCC diagnosis, ultimately improving patient care.
Keywords: cutaneous squamous cell carcinoma , deep learning , histological grading , transfer learning , artificial intelligence
Introduction
Cutaneous squamous cell carcinoma (cSCC), a prevalent skin cancer originating from the outer skin layer, arises from abnormal cell growth (dysplasia) that progresses to invasive cancer [1, 2]. It commonly affects sun-exposed areas, with a predilection for the head and neck [2, 3]. Incidence rates are increasing globally, particularly in Western countries, with higher rates observed in older individuals and men [4, 5, 6, 7, 8].
While generally less aggressive than melanoma, cSCC can be lethal, especially if diagnosed late or with certain high-risk subtypes [9, 10]. The global rise in cSCC cases underscores the need for comprehensive public health strategies, including prevention, early detection, and effective treatment [5, 8].
Histological grading, based on the degree of cellular differentiation, is crucial for prognosis and treatment decisions [2, 11]. It guides interventions, ranging from surgical excision for low-grade tumors to Mohs surgery, radiation therapy, or systemic treatments like Cemiplimab for high-grade or advanced cases [12, 13]. Furthermore, histological grading aids in identifying high-risk patients who require closer monitoring and facilitates research into the molecular underpinnings of cSCC progression [13, 14, 15].
Artificial intelligence (AI), as part of machine learning (ML) and deep learning (DL) in particular, is transforming histopathology by enhancing diagnostic accuracy (ACC), prognostication, and disease management [16, 17, 18]. ML algorithms, trained on digitized tissue slides, can identify patterns indicative of specific pathologies, augmenting pathologists’ capabilities and potentially leading to faster, more precise diagnoses [19, 20, 21].
DL, especially convolutional neural networks (CNNs), has emerged as a powerful tool for histopathology image analysis, facilitating tasks like tissue classification, tumor detection, and biomarker identification [22, 23, 24]. Moreover, AI models can predict disease progression and treatment responses, enabling personalized medicine approaches [18, 25].
Transfer learning (TL), a technique where a pre-trained model is adapted to a new task, is particularly valuable in histopathology due to the limited availability of annotated data [26]. By leveraging knowledge from models trained on large, general datasets like ImageNet, TL significantly reduces the need for extensive training data and accelerates model development while often enhancing performance on specific histopathology tasks [22, 27].
As shown above, histological grading of cSCC serves as a critical prognostic indicator, significantly influencing therapeutic decision-making and risk stratification, which are essential for optimal clinical management [10, 11, 12, 13, 28]. However, the current reliance on manual grading introduces inherent subjectivity, leading to possible interobserver variability due to differences in pathologists’ experience, training, and potential fatigue. Furthermore, manual grading is time-consuming, particularly with complex specimens, and suffers from limited reproducibility due to the potential for subjective interpretation of grading criteria. Additionally, the process necessitates trained pathologists, a resource that may be scarce in certain settings.
Aim
Given the inherent limitations of manual grading in cSCC, the aim of this study was to develop and validate a DL-based model for automated cSCC grading that can objectively and consistently classify cSCC in three grading classes (G1, G2, and G3), potentially reducing interobserver variability, improving diagnostic efficiency and time, and ultimately enhancing patient care.
Materials and Methods
Materials
Forty-seven consecutive cases of cSCC, obtained with approval from the Research Ethics Committees of the Emergency Clinical County Hospital, Cluj-Napoca, Romania (Approval No. 38789/13.09.2021) and the Iuliu Haţieganu University of Medicine and Pharmacy, Cluj-Napoca (AVZ2/8.11.2021), were included in this study. Following standard histological processing and Hematoxylin–Eosin (HE) staining, the excised tissue sections were scanned using a Pannoramic SCAN II slide scanner (3DHISTECH, Budapest, Hungary) with a ×40 objective. From each whole-slide image (WSI), 1000×1000 pixel crops in a 24-bit red, green, and blue (RGB) color space were extracted by a trained pathologist. A panel of three pathologists independently reviewed these crops, and only those with unanimous diagnostic agreement were retained for subsequent analysis.
For each of the three grades (G1, G2, and G3), 100 images were selected. Samples of the selected images for each grade are presented in Figure 1.
Figure 1.
Dataset sample. The first row resembles images labeled as G1, the second row as G2, and the third row as G3. Hematoxylin–Eosin (HE) staining, ×200
Methods
We trained three distinct classification deep neural networks (DNNs) with diverse architectures, originally designed for classifying real-world images into 1000 categories from the ImageNet dataset [29]. Utilizing the concept of TL [30], we adapted these networks to classify SCC images into three grading scores: G1, G2, and G3. The networks underwent modification, re-training, and thorough statistical evaluation. The top-performing network from each architecture were tested on 60 new images, with results compared to assessments by a panel of three pathologists. All computational tasks were conducted using MATLAB® (MathWorks, USA).
The three DNNs were selected based on previous research [31, 32] and their balance between ACC and prediction time [33]. The first network, AlexNet [34], is a CNN comprising five convolutional layers interspersed with max-pooling layers, followed by two fully connected layers, and concluding with a 1000-class Softmax layer. This model, introduced in 2012, can be freely downloaded [35] and is available in MATLAB. The second model, GoogLeNet [36], features a 22-layer architecture with the innovative “inception” modules, significantly different from AlexNet. Released in 2014, this model is also freely available [37] and accessible in MATLAB. The third model, ResNet-18 [38], incorporates a non-linear, 18-layer architecture known for its residual learning approach, introduced in 2016. This model is likewise available in MATLAB.
For each model, the final classification layers were replaced to adapt to the new task of classifying SCC images into three grading classes. Specifically, the last fully connected layer and the output layer were substituted to align with our dataset’s labels.
The modified networks were re-trained on the new dataset, with 10% of images reserved for validation and 90% used for training, employing a 10-fold cross-validation method. Training hyperparameters were standardized across all architectures, with a mini-batch size of 100, an initial learning rate of 0.0001, and a validation patience of four epochs. Stochastic gradient descent with momentum was utilized as the optimizer. Performance metrics included mean ACC, sensitivity (SN), specificity (SP), and area under the curve (AUC).
Clinical evaluation
A set of 60 images from six consecutive SCC cases were randomly selected from new WSIs, following the same procedure as the original dataset creation. These images were classified using the application and independently evaluated by three pathologists. Agreement between the networks and the pathologists was assessed.
Statistical assessment
Due to the stochastic nature of DNN algorithms, which are influenced by the specific training and validation datasets, we ensured robust results by running each model 100 times using 10-fold cross-validation. For each run, training and validation data were randomly selected but kept consistent across all three networks to enable direct performance comparisons. The normality of the data distribution was assessed using the Kolmogorov–Smirnov, Lilliefors, and Shapiro–Wilk tests.
The agreement between the networks and the panel of pathologists was evaluated using Cohen’s kappa (κ) coefficient [39]. The coefficient ranges between 0 and 1, with values close to 0 indicating no agreement and values close to 1 indicating perfect agreement. For interpretability and consistency, Landis & Koch [40] guidelines were used.
Results
A total of 300 trained networks resulted, 100 having AlexNet architecture, 100 having GoogLeNet architecture, and the last 100 having ResNet-18 architecture. Mean and standard deviation (SD) for ACC, SN, SP, and AUC of the networks, on the hole dataset are presented in Table 1.
Table 1.
Performance assessment, mean ± SD
|
AlexNet |
GoogLeNet |
ResNet-18 |
ANOVA |
|
|
ACC [%] |
85.5±5.5 |
86.2±6.07 |
85.5±5.6 |
p >0.05 |
|
SN |
0.854±0.055 |
0.862±0.060 |
0.855±0.056 |
p >0.05 |
|
SP |
0.927±0.027 |
0.931±0.030 |
0.927±0.028 |
p >0.05 |
|
AUC |
0.983±0.016* |
0.978±0.021 |
0.973±0.023* |
p <0.05 |
ACC: Accuracy; ANOVA: Analysis of variance; AUC: Area under the curve; SD: Standard deviation; SN: Sensitivity; SP: Specificity. *Student’s t-test, p<0.05
All normality tests (Kolmogorov–Smirnov, Lilliefors, and Shapiro–Wilk) indicated that the performance assessment variables did not follow a normal distribution. Nevertheless, in accordance with the central limit theorem [41], and given a sample size of 100, parametric tests were considered for comparing means.
One-way analysis of variance (ANOVA) showed that there are no significant differences between the performance parameters of the three networks, except for the AUC. Student’s t-test confirms there is a significant difference between the AUC of AlexNet and ResNet-18.
Confusion matrices on all dataset images from the best performing networks of each architecture are presented in Tables 2, 3, 4.
Table 2.
AlexNet confusion matrix
| Predicted class |
G1 |
G2 |
G3 |
|
G1 |
99 |
1 |
0 |
|
G2 |
2 |
97 |
1 |
|
G3 |
0 |
1 |
100 |
|
Target class |
|||
Table 3.
GoogLeNet confusion matrix
| Predicted class |
G1 |
G2 |
G3 |
|
|
G1 |
99 |
1 |
0 |
|
|
G2 |
4 |
95 |
1 |
|
|
G3 |
1 |
1 |
98 |
|
|
Target class |
||||
Table 4.
ResNet-18 confusion matrix
| Predicted class |
G1 |
G2 |
G3 |
|
G1 |
98 |
2 |
0 |
|
G2 |
0 |
99 |
1 |
|
G3 |
9 |
1 |
99 |
|
Target class |
|||
The misclassification patterns are evident from the confusion matrices presented in Tables 2, 3, 4. These tables show that G1 was frequently misclassified as G2, and G3 was also often misclassified as G2. Conversely, G2 was misclassified as either G1 or G3. Misclassifications between G1 and G3 were very rare.
A sample of misclassified images is shown in Figure 2A, 2B, 2C, 2D, 2E. As anticipated, most G1 images were misclassified as G2. Similarly, G2 images were often incorrectly classified as G1 or G3, whereas G3 images were only misclassified as G2.
Figure 2.
Misclassified images: (A) A G1 pattern misclassified as G2; (B and C) A G2 pattern misclassified as G1; (D) A G2 pattern misclassified as G3; (E) A G3 pattern misclassified as G2. HE staining, ×200
The best performing networks from each architecture were packed as MATLAB files and are offered on a public repository (https://github.com/Mircea-Sebastian-Serbanescu/Automated-Cutaneous-Squamous-Cell-Carcinoma-Grading-Using-Deep-Learning-with-Transfer-Learning.git).
All Cohen’s kappa were above 0.8 (Table 5), showing almost perfect agreement and concluding that the observed agreement is not accidental. The average agreement between the pathologists (0.8889) exceeded the average agreement between the networks (0.8722). However, both are very high. Panel pathologist P3 had the best agreement with ResNet-18 close to 1 (0.9667).
Table 5.
Clinical evaluation assessment: Cohen’s kappa (κ) coefficient. P1–P3 represent the three panel pathologists. Sixty images from six different cases
|
AlexNet |
GoogLeNet |
ResNet-18 |
P1 |
P2 |
P3 |
|
|
AlexNet |
0.8833 |
0.9000 |
0.9000 |
0.8833 |
0.9000 |
|
|
GoogLeNet |
0.8333 |
0.8667 |
0.9167 |
0.8667 |
||
|
ResNet-18 |
0.9333 |
0.8667 |
0.9667 |
|||
|
P1 |
0.8667 |
0.9333 |
||||
|
P2 |
0.8667 |
Discussions
AI has emerged as a valuable tool across various fields of medicine [42]. In radiology, AI algorithms can analyze medical images to detect and diagnose conditions like cancer and fractures with remarkable ACC [43, 44]. AI-powered tools in pathology can help identify abnormalities in tissue samples, aiding in disease diagnosis and prognosis [24]. In drug discovery, AI models have accelerated the identification of potential therapeutic targets and the development of new medications [45]. Furthermore, AI has shown promise in predicting patient outcomes and personalizing treatment plans based on individual characteristics and disease profiles [46].
The application of DL in pathology was delayed compared to other medical fields due to the necessity of digital slide scanners. These scanners were crucial for converting physical slides into digital data, which is essential for training DL models [26]. The availability of large datasets, coupled with advancements in computational power and algorithm development, finally allowed DL to emerge as a powerful tool in pathology, enabling significant progress in areas such as cancer diagnosis and prognosis [47].
Our research has demonstrated robust results in the grade scoring of cSCC through the application of TL techniques and general-purpose DNNs. The models achieved mean ACCs exceeding 0.85, SN of 0.85, SP of 0.92, and an AUC greater than 0.97. A statistical evaluation of the three networks employed in this study revealed comparable performance outcomes, except for a slight disparity in AUC between AlexNet and ResNet-18.
Furthermore, clinical evaluations showed a high degree of agreement between the network predictions and the assessments made by pathologists. There was also strong concordance among the pathologists themselves, as well as between different network architectures. These findings are consistent with the performance metrics obtained during the classification performance assessment phase, underscoring the possible reliability and ACC of the DL models in clinical applications and indicating that the grading features of cSCC are well-separated.
Analyzing the misclassified images shown in Figure 2A, 2B, 2C, 2D, 2E, we can intuitively understand the reasons behind the network’s decisions. The G1 pattern was misclassified as G2, likely due to the dense inflammatory infiltrate and potentially processing artifacts, such as the detachment of keratin pearls during sectioning. The G2 pattern was mistaken for G1, probably because no staining normalization or assessment was performed, leading to differences in staining. In hypereosinophilic slides, the network may misinterpret clusters of larger cells with pyknotic nuclei arranged in a whorl-like pattern as keratin pearls. Inflammation might also explain why the G2 pattern was misclassified as G3. Additionally, the small size of the keratin pearls could also contribute to this misinterpretation. Finally, the G3 pattern was misclassified as G2, likely because some necrosis resembled keratin formations. Furthermore, the presence of intervening stroma within the neoplastic cell sheets, particularly those containing blood vessels with brightly eosinophilic erythrocytes or sites of hemorrhage, may have contributed to the network’s misidentification of these areas as keratin pearls.
The DNNs have demonstrated strong classification capabilities across various medical imaging fields and pathologies. Our previous research has highlighted their effectiveness, including automatic segmentation of liver lesions in ultrasound (US) investigations [48] and the histological class prediction of breast tumors in mammography [31]. Achieving reliable results in the histopathological domain further underscores the potential utility of other imaging techniques, such as high-frequency US, which is an essential non-invasive tool for the pre-therapeutic assessment of basal cell carcinoma (BCC) [49].
These three network architectures – AlexNet, GoogLeNet, and ResNet-18 – have been previously utilized in the classification of BCC patterns [26]. While earlier studies reported good performance, the current approach has yielded superior results. Several factors may contribute to this improvement. Notably, the use of a balanced dataset likely enhanced the model’s performance. Additionally, the current task involved classifying images into only three categories, which may be inherently simpler. The grading task itself may also be more “clearly separable” with better inter-human agreement reinforcing this hypothesis. Comparable outcomes have been observed in automated Gleason grading of prostate cancer [32, 50], where TL from general-purpose DNNs achieved notable results. However, the performance in these studies was slightly lower compared to the current research.
Furthermore, while other malignancies often require costly investigations for reliable outcome prediction, such as immunohistochemistry predictive markers for primary colorectal cancer tumors [51], this is not the case for cSCC [2], making the current approach both efficient and safe.
The predominant application of DL in the context of cSCC research, as evidenced by existing literature, centers on the utilization of clinical and dermatoscopic imaging modalities. Studies conducted by Han et al. [52], Jain et al. [53], Bechelli & Delhommelle [54], Fraiwan & Faouri [55], Bassel et al. [56], Zia Ur Rehman et al. [57], and Hussain et al. [58] have employed different CNNs architectures and ML techniques to achieve automated diagnosis. The studies researched either the differentiation of benign from malignant skin lesions, including images of cSCC, or the classification of specific subtypes of skin malignancies, based on clinical or dermatoscopic image analysis. In addition to lesion classification, researchers such as Zhao et al. [59] have explored the use of DL models to assess the risk stratification of skin tumors, categorizing them into low-risk, high-risk, and dangerous categories.
While research utilizing DL on digital histopathology images of cSCC remains limited, Thomas et al. [60] developed an interpretable DL model for multi-class segmentation and classification of non-melanoma skin cancers from histopathology images. They used 290 slides of cSCC, and intraepidermal carcinoma, captured at 10× magnification with a resolution of 0.67 μm/pixel. The model, based on a U-Net architecture with a ResNet-50 encoder, achieved high ACC in both pixel-level segmentation (85%) and whole-image classification (93.6%). The study demonstrated the feasibility of using DL for automated skin cancer diagnosis and the potential for such models to assist pathologists in tasks like surgical margin assessment.
Wako et al. [61] developed a DL system to classify SCC margins in histopathology images. Using 3972 images for training, 492 for validation, and 84 for testing, they employed pre-trained models VGG16, ResNet-50, MobileNetV2, and EfficientNetB0, achieving 95.2% ACC [61]. This demonstrates the potential of DL to improve the ACC and efficiency of SCC margin assessment [61].
Davis et al. employed DL on 95 WSIs of frozen sections obtained from Mohs micrographic surgery for cSCC [62]. Utilizing 256×256-pixel image patches and a ResNet-101 model for tumor classification at the patch level, the algorithm achieved an AUC of 98.1%. Notably, performance was superior for poorly and moderately differentiated tumors (AUC 96.8%) compared to well-differentiated tumors (AUC 89.5%).
Ruini et al. developed a ML model using MobileNetV2 to automatically detect SCC in ex vivo confocal laser scanning microscopy (CLSM) images [63]. Trained and validated on 543 grayscale CLSM images, the model achieved a high AUC of 0.95, demonstrating the potential for faster and more accurate SCC diagnosis [633]. While promising, further validation with external datasets is needed to confirm its generalizability.
In a retrospective analysis involving 104 cSCC patients, Pérez-Baena et al. [64] developed a novel ML-derived cellular morphometric risk score (CMRS) and compared it to the Brigham and Women’s Hospital (BWH) staging system. The CMRS demonstrated independent prognostic value, but when combined with the BWH staging system, it significantly enhanced the ACC of risk stratification, especially regarding local recurrence and nodal metastasis. This approach objectively underscores the added value of ML in clinical prognostication.
Study limitations
A significant limitation of this study is the size of the dataset. Large datasets are essential for optimizing the performance of image analysis systems. Our dataset should be considered small for this specific task. Despite this, we have mitigated the issue of class imbalance, a common challenge in medical classification tasks, through the study’s design. This approach allowed us to achieve better than expected results.
Another limitation is that the data originates from a single medical center and was scanned using a single WSI scanner. More than that, the staining was not normalized [65].
Stroma and deep learning: a future direction
The tumor microenvironment (TME) plays a crucial role in cancer progression, and the stroma, a major component of the TME, has been implicated in cSCC development and metastasis [15]. The stroma consists of various cell types, including fibroblasts, immune cells, and endothelial cells, as well as extracellular matrix (ECM) components. Interactions between tumor cells and stromal cells can promote tumor growth, angiogenesis, and immune evasion [66]. The composition of the stroma, including the presence of specific immune cells and the organization of the ECM, can influence the aggressiveness of cSCC and its response to therapy [64].
DL models can be leveraged to analyze the stromal compartment of cSCC, potentially revealing novel insights into tumor–stroma interactions. For instance, DL algorithms could be trained to quantify stromal features such as fibroblast density, collagen deposition, vessel density and position and immune cell infiltration. These features could then be correlated with clinical outcomes, such as tumor aggressiveness and metastatic potential, to identify prognostic biomarkers and potential therapeutic targets.
Furthermore, DL models could be used to analyze spatial relationships between tumor cells and stromal components. This could help elucidate the mechanisms by which stromal cells promote tumor growth and metastasis, paving the way for the development of novel therapeutic strategies targeting the TME. For example, DL could be used to identify specific patterns of stromal cell infiltration that are associated with poor prognosis or resistance to therapy.
Conclusions
This paper presents an approach for cSCC grading using image classification based on TL in DL, trained on a dataset of only 300 images. DNN derived from AlexNet, GoogLeNet, and ResNet-18 were used, achieving average ACC, SN, SP, and AUC scores of approximately 0.85, 0.85, 0.93, and 0.97, respectively. For this task, all networks performed similarly, indicating that the grading features of cSCC are well-separated. In the clinical assessment, grading agreement measured through Cohen’s kappa showed that the neural networks had similar agreement levels with the pathologists and among themselves. There is also a very high agreement between pathologists on cSCC grading. The top-performing networks from each architecture have been made available in a public repository.
Conflict of interests
The authors declare that they have no conflict of interests.
References
- 1.Caudill J, Thomas JE, Burkhart CG. The risk of metastases from squamous cell carcinoma of the skin. Int J Dermatol. 2023;62(4):483–486. doi: 10.1111/ijd.16164. [DOI] [PubMed] [Google Scholar]
- 2. Scolyer RA , et al., editors. WHO Classification of Tumours Editorial Board . Skin tumours . 5th . Lyon, France : International Agency for Research on Cancer (IARC) Press ; 2023 . Keratinocytic/epidermal tumours . [Google Scholar]
- 3.Fijałkowska M, Koziej M, Antoszewski B. Detailed head localization and incidence of skin cancers. Sci Rep. 2021;11(1):12391–12391. doi: 10.1038/s41598-021-91942-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Tokez S, Wakkee M, Louwman M, Noels E, Nijsten T, Hollestein L. Assessment of cutaneous squamous cell carcinoma (cSCC) in situ incidence and the risk of developing invasive cSCC in patients with prior cSCC in situ vs the general population in the Netherlands, 1989-2017. JAMA Dermatol. 2020;156(9):973–981. doi: 10.1001/jamadermatol.2020.1988. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Huang S, Jiang J, Wong HS, Zhu P, Ji X, Wang D. Global burden and prediction study of cutaneous squamous cell carcinoma from 1990 to 2030: a systematic analysis and comparison with China. J Glob Health. 2024;14:04093–04093. doi: 10.7189/jogh.14.04093. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Eggermont CJ, Eggermont AMM. Shifting landscape in skin cancer incidence: the rising tide of cutaneous squamous cell carcinoma and potential implications for prevention. Br J Dermatol. 2024;190(4):460–461. doi: 10.1093/bjd/ljad480. [DOI] [PubMed] [Google Scholar]
- 7.Leiter U, Keim U, Garbe C. Epidemiology of skin cancer: update 2019. Adv Exp Med Biol. 2020;1268:123–139. doi: 10.1007/978-3-030-46227-7_6. [DOI] [PubMed] [Google Scholar]
- 8.Guo A, Liu X, Li H, Cheng W, Song Y. The global, regional, national burden of cutaneous squamous cell carcinoma (1990-2019) and predictions to 2035. Eur J Cancer Care. 2023 doi: 10.1111/ecc.13689. [DOI] [PubMed] [Google Scholar]
- 9.Migden MR, Rischin D, Schmults CD, Guminski A, Hauschild A, Lewis KD, Chung CH, Hernandez-Aya L, Lim AM, Chang ALS, Rabinowits G, Thai AA, Dunn LA, Hughes BGM, Khushalani NI, Modi B, Schadendorf D, Gao B, Seebach F, Li S, Li J, Mathias M, Booth J, Mohan K, Stankevich E, Babiker HM, Brana I, Gil-Martin M, Homsi J, Johnson ML, Moreno V, Niu J, Owonikoko TK, Papadopoulos KP, Yancopoulos GD, Lowy I, Fury MG. PD-1 blockade with Cemiplimab in advanced cutaneous squamous-cell carcinoma. N Engl J Med. 2018;379(4):341–351. doi: 10.1056/NEJMoa1805131. [DOI] [PubMed] [Google Scholar]
- 10.Martorell-Calatayud A, Sanmartín Jimenez, Cruz Mojarrieta, Guillén Barona. Cutaneous squamous cell carcinoma: defining the high-risk variant. Actas Dermosifiliogr. 2013;104(5):367–379. doi: 10.1016/j.adengl.2011.12.012. [DOI] [PubMed] [Google Scholar]
- 11.Schmitz L, Kanitakis J. Histological classification of cutaneous squamous cell carcinomas with different severity. J Eur Acad Dermatol Venereol. 2019;33(Suppl 8):11–15. doi: 10.1111/jdv.15950. [DOI] [PubMed] [Google Scholar]
- 12.Maubec E. Update of the management of cutaneous squamous-cell carcinoma. Acta Derm Venereol. 2020;100(11):adv00143–adv00143. doi: 10.2340/00015555-3498. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.García-Foncillas J, Tejera-Vaquerizo A, Sanmartín O, Rojo F, Mestre J, Martín S, Azinovic I, Mesía R. Update on management recommendations for advanced cutaneous squamous cell carcinoma. Cancers (Basel) 2022;14(3):629–629. doi: 10.3390/cancers14030629. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Feng J, Zhang W, Zeng W, Dong X, Wang Y, Gu Y, Lan Y, Yang W, Lu H. Expression analysis of retinal G protein-coupled receptor and its correlation with regulation of the balance between proliferation and aberrant differentiation in cutaneous squamous cell carcinoma. Acta Derm Venereol. 2024;104:adv13213–adv13213. doi: 10.2340/actadv.v104.13213. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Fania L, Didona D, Di Pietro, Verkhovskaia S, Morese R, Paolino G, Donati M, Ricci F, Coco V, Ricci F, Candi E, Abeni D, Dellambra E. Cutaneous squamous cell carcinoma: from pathophysiology to novel therapeutic approaches. Biomedicines. 2021;9(2):171–171. doi: 10.3390/biomedicines9020171. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Ahmad Z, Rahim S, Zubair M, Abdul-Ghafar J. Artificial intelligence (AI) in medicine, current applications and future role with special emphasis on its potential and promise in pathology: present and future impact, obstacles including costs and acceptance among pathologists, practical and philosophical considerations. A comprehensive review. Diagn Pathol. 2021;16(1):24–24. doi: 10.1186/s13000-021-01085-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Coudray N, Ocampo PS, Sakellaropoulos T, Narula N, Snuderl M, Fenyö D, Moreira AL, Razavian N, Tsirigos A. Classification and mutation prediction from non-small cell lung cancer histopathology images using deep learning. Nat Med. 2018;24(10):1559–1567. doi: 10.1038/s41591-018-0177-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Kather JN, Krisam J, Charoentong P, Luedde T, Herpel E, Weis CA, Gaiser T, Marx A, Valous NA, Ferber D, Jansen L, Reyes-Aldasoro CC, Zörnig I, Jäger D, Brenner H, Chang-Claude J, Hoffmeister M, Halama N. Predicting survival from colorectal cancer histology slides using deep learning: a retrospective multicenter study. PLoS Med. 2019;16(1):e1002730–e1002730. doi: 10.1371/journal.pmed.1002730. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Shafi S, Parwani AV. Artificial intelligence in diagnostic pathology. Diagn Pathol. 2023;18(1):109–109. doi: 10.1186/s13000-023-01375-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Obermeyer Z, Emanuel EJ. Predicting the future - big data, machine learning, and clinical medicine. N Engl J Med. 2016;375(13):1216–1219. doi: 10.1056/NEJMp1606181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Marini N, Marchesin S, Otálora S, Wodzinski M, Caputo A, van Rijthoven, Aswolinskiy W, Bokhorst JM, Podareanu D, Petters E, Boytcheva S, Buttafuoco G, Vatrano S, Fraggetta F, van der, Agosti M, Ciompi F, Silvello G, Muller H, Atzori M. Unleashing the potential of digital pathology data by training computer-aided diagnosis models without human annotations. NPJ Digit Med. 2022;5(1):102–102. doi: 10.1038/s41746-022-00635-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, van der, van Ginneken, Sánchez CI. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 23.Manakitsa N, Maraslidis GS, Moysis L, Fragulis GF. A review of machine learning and deep learning for object detection, semantic segmentation, and human action recognition in machine and robotic vision. Technologies. 2024;12(2):15–15. [Google Scholar]
- 24.Bera K, Schalper KA, Rimm DL, Velcheti V, Madabhushi A. Artificial intelligence in digital pathology - new tools for diagnosis and precision oncology. Nat Rev Clin Oncol. 2019;16(11):703–715. doi: 10.1038/s41571-019-0252-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Janowczyk A, Madabhushi A. Deep learning for digital pathology image analysis: a comprehensive tutorial with selected use cases. J Pathol Inform. 2016;7:29–29. doi: 10.4103/2153-3539.186902. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Bungărdean RM, Şerbănescu MS, Streba CT, Crişan M. Deep learning with transfer learning in pathology. Case study: classification of basal cell carcinoma. Rom J Morphol Embryol. 2021;62(4):1017–1028. doi: 10.47162/RJME.62.4.14. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Cheplygina V, de Bruijne, Pluim JPW. Not-so-supervised: a survey of semi-supervised, multi-instance, and transfer learning in medical image analysis. Med Image Anal. 2019;54:280–296. doi: 10.1016/j.media.2019.03.009. [DOI] [PubMed] [Google Scholar]
- 28.Kabir S, Schmults CD, Ruiz ES. A review of cutaneous squamous cell carcinoma epidemiology, diagnosis, and management. Int J Cancer Manag. 2018;11(1):e60846–e60846. [Google Scholar]
- 29.ImageNet, 2021. ImageNet [online] 2024 Available from: https://image-net.org/
- 30.Bozinovski S. Reminder of the first paper on transfer learning in neural networks, 1976. Informatica. 2020;44(3):291–302. [Google Scholar]
- 31.Nica RE, Şerbănescu MS, Florescu LM, Camen GC, Streba CT, Gheonea IA. Deep learning: a promising method for histological class prediction of breast tumors in mammography. J Digit Imaging. 2021;34(5):1190–1198. doi: 10.1007/s10278-021-00508-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.erbănescu MS, Manea NC, Streba L, Belciug S, Pleşea IE, Pirici I, Bungărdean RM, Pleşea RM. Automated Gleason grading of prostate cancer using transfer learning from general-purpose deep-learning networks. Rom J Morphol Embryol. 2020;61(1):149–155. doi: 10.47162/RJME.61.1.17. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.MATLAB, 2024: Pretrained deep neural networks. MATLAB & Simulink [online] 2024 Available from: https://www.mathworks.com/help/deeplearning/ug/pretrained-convolutional-neural-networks.html.
- 34. Krizhevsky A , Sutskever I , Hinton G . In: Advances in neural information processing systems (NIPS) 25. 26th Annual Conference on NIPS 2012 . Pereira F , et al., editors. New York, USA : Curran Associates, Inc. ; 2013 . ImageNet classification with deep convolutional neural networks ; pp. 1097 – 1105 . [Google Scholar]
- 35.GitHub, 2021: BVLC AlexNet model [online] 2024 Available from: https://github.com/BVLC/caffe/tree/master/models/bvlc_alexnet.
- 36.Szegedy C, Liu W, Jia Y, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A. Going deeper with convolutions . 2015 Institute of Electrical and Electronics Engineers (IEEE) Conference on Computer Vision and Pattern Recognition (CVPR); 2015 Jun 7-12 ; Boston, MA, USA. [Google Scholar]
- 37.GitHub, 2021: BVLC GoogLeNet model [online] 2024 Available from: https://github.com/BVLC/caffe/tree/master/models/bvlc_googlenet.
- 38.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition . 2016 Institute of Electrical and Electronics Engineers (IEEE) Conference on Computer Vision and Pattern Recognition (CVPR) ; 2016 Jun 27-30 ; Las Vegas, NV, USA. [Google Scholar]
- 39.McHugh ML. Interrater reliability: the kappa statistic. Biochem Med (Zagreb) 2012;22(3):276–282. [PMC free article] [PubMed] [Google Scholar]
- 40.Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33(1):159–174. [PubMed] [Google Scholar]
- 41. Montgomery DC , Runger GC . Applied statistics and probability for engineers . 6th . Hoboken, NJ, USA : John Wiley & Sons, Inc. ; 2014 . [Google Scholar]
- 42.Bekbolatova M, Mayer J, Ong CW, Toma M. Transformative potential of AI in healthcare: definitions, applications, and navigating the ethical landscape and public perspectives. Healthcare (Basel) 2024;12(2):125–125. doi: 10.3390/healthcare12020125. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Bi WL, Hosny A, Schabath MB, Giger ML, Birkbak NJ, Mehrtash A, Allison T, Arnaout O, Abbosh C, Dunn IF, Mak RH, Tamimi RM, Tempany CM, Swanton C, Hoffmann U, Schwartz LH, Gillies RJ, Huang RY, Aerts HJWL. Artificial intelligence in cancer imaging: clinical challenges and applications. CA Cancer J Clin. 2019;69(2):127–157. doi: 10.3322/caac.21552. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts HJWL. Artificial intelligence in radiology. Nat Rev Cancer. 2018;18(8):500–510. doi: 10.1038/s41568-018-0016-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Paul D, Sanap G, Shenoy S, Kalyane D, Kalia K, Tekade RK. Artificial intelligence in drug discovery and development. Drug Discov Today. 2021;26(1):80–93. doi: 10.1016/j.drudis.2020.10.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Hilton CB, Milinovich A, Felix C, Vakharia N, Crone T, Donovan C, Proctor A, Nazha A. Personalized predictions of patient outcomes during and after hospitalization using artificial intelligence. NPJ Digit Med. 2020;3:51–51. doi: 10.1038/s41746-020-0249-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Tran KA, Kondrashova O, Bradley A, Williams ED, Pearson JV, Waddell N. Deep learning in cancer diagnosis, prognosis and treatment selection. Genome Med. 2021;13(1):152–152. doi: 10.1186/s13073-021-00968-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Florescu LM, Streba CT, Şerbănescu MS, Mămuleanu M, Florescu DN, Teică RV, Nica RE, Gheonea IA. Federated learning approach with pre-trained deep learning models for COVID-19 detection from unsegmented CT images. Life (Basel) 2022;12(7):958–958. doi: 10.3390/life12070958. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Bungărdean RM, Şerbănescu MS, Colosi HA, Crişan M. High-frequency ultrasound: an essential non-invasive tool for the pre-therapeutic assessment of basal cell carcinoma. Rom J Morphol Embryol. 2021;62(2):545–551. doi: 10.47162/RJME.62.2.21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.erbănescu MS, Oancea CN, Streba CT, Pleşea IE, Pirici D, Streba L, Pleşea RM. Agreement of two pre-trained deep learning neural networks built with transfer learning with six pathologists on 6000 patches of prostate cancer from Gleason2019 Challenge. Rom J Morphol Embryol. 2020;61(2):513–519. doi: 10.47162/RJME.61.2.21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Bărbălan A, Nicolaescu AC, Măgăran AV, Mercuţ R, Bălăşoiu M, Băncescu G, Şerbănescu MS, Lazăr OF, Săftoiu A. Immunohistochemistry predictive markers for primary colorectal cancer tumors: where are we and where are we going. Rom J Morphol Embryol. 2018;59(1):29–42. [PubMed] [Google Scholar]
- 52.Han SS, Kim MS, Lim W, Park GH, Park I, Chang SE. Classification of the clinical images for benign and malignant cutaneous tumors using a deep learning algorithm. J Invest Dermatol. 2018;138(7):1529–1538. doi: 10.1016/j.jid.2018.01.028. [DOI] [PubMed] [Google Scholar]
- 53.Jain S, Singhania U, Tripathy B, Nasr EA, Aboudaif MK, Kamrani AK. Deep learning-based transfer learning for classification of skin cancer. Sensors (Basel) 2021;21(23):8142–8142. doi: 10.3390/s21238142. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Bechelli S, Delhommelle J. Machine learning and deep learning algorithms for skin cancer classification from dermoscopic images. Bioengineering (Basel) 2022;9(3):97–97. doi: 10.3390/bioengineering9030097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Fraiwan M, Faouri E. On the automatic detection and classification of skin cancer using deep transfer learning. Sensors (Basel) 2022;22(13):4963–4963. doi: 10.3390/s22134963. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Bassel A, Abdulkareem AB, Alyasseri ZAA, Sani NS, Mohammed HJ. Automatic malignant and benign skin cancer classification using a hybrid deep learning approach. Diagnostics (Basel) 2022;12(10):2472–2472. doi: 10.3390/diagnostics12102472. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Zia Ur, Ahmed F, Alsuhibany SA, Jamal SS, Zulfiqar Ali, Ahmad J. Classification of skin cancer lesions using explainable deep learning. Sensors (Basel) 2022;22(18):6915–6915. doi: 10.3390/s22186915. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Hussain M, Khan MA, Damaševičius R, Alasiry A, Marzougui M, Alhaisoni M, Masood A. SkinNet-INIO: multiclass skin lesion localization and classification using fusion-assisted deep neural networks and improved nature-inspired optimization algorithm. Diagnostics (Basel) 2023;13(18):2869–2869. doi: 10.3390/diagnostics13182869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Zhao XY, Wu X, Li FF, Li Y, Huang WH, Huang K, He XY, Fan W, Wu Z, Chen ML, Li J, Luo ZL, Su J, Xie B, Zhao S. The application of deep learning in the risk grading of skin tumors for patients using clinical images. J Med Syst. 2019;43(8):283–283. doi: 10.1007/s10916-019-1414-2. [DOI] [PubMed] [Google Scholar]
- 60.Thomas SM, Lefevre JG, Baxter G, Hamilton NA. Interpretable deep learning systems for multi-class segmentation and classification of non-melanoma skin cancer. Med Image Anal. 2021;68:101915–101915. doi: 10.1016/j.media.2020.101915. [DOI] [PubMed] [Google Scholar]
- 61.Wako BD, Dese K, Ulfata RE, Nigatu TA, Turunbedu SK, Kwa T. Squamous cell carcinoma of skin cancer margin classification from digital histopathology images using deep learning. Cancer Control. 2022;29:10732748221132528–10732748221132528. doi: 10.1177/10732748221132528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Davis MJ, Srinivasan G, Chacko R, Chen S, Suvarna A, Vaickus LJ, Torres VC, Hodge S, Chen EY, Preum S, Samkoe KS, Christensen BC, LeBoeuf MR, Levy JJ. A deep learning algorithm to detect cutaneous squamous cell carcinoma on frozen sections in Mohs micrographic surgery: a retrospective assessment. Exp Dermatol. 2024;33(1):e14949–e14949. doi: 10.1111/exd.14949. [DOI] [PubMed] [Google Scholar]
- 63.Ruini C, Schlingmann S, Jonke Ž, Avci P, Padrón-Laso V, Neumeier F, Koveshazi I, Ikeliani IU, Patzer K, Kunrad E, Kendziora B, Sattler E, French LE, Hartmann D. Machine learning based prediction of squamous cell carcinoma in ex vivo confocal laser scanning microscopy. Cancers (Basel) 2021;13(21):5522–5522. doi: 10.3390/cancers13215522. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Pérez-Baena MJ, Mao JH, Pérez-Losada J, Santos-Briz Á, Chang H, Cañueto J. Artificial intelligence-empowered cellular morphometric risk score improves prognostic stratification of cutaneous squamous cell carcinoma. Clin Exp Dermatol. 2024;49(7):692–698. doi: 10.1093/ced/llad264. [DOI] [PubMed] [Google Scholar]
- 65.erbănescu MS, Pleşea IE. A hardware approach for histological and histopathological digital image stain normalization. Rom J Morphol Embryol. 2015;56(2 Suppl):735–741. [PubMed] [Google Scholar]
- 66.Buruiană A, Gheban BA, Gheban-Roşca IA, Georgiu C, Crişan D, Crişan M. The tumor stroma of squamous cell carcinoma: a complex environment that fuels cancer progression. Cancers (Basel) 2024;16(9):1727–1727. doi: 10.3390/cancers16091727. [DOI] [PMC free article] [PubMed] [Google Scholar]


