Skip to main content
Clinical and Translational Radiation Oncology logoLink to Clinical and Translational Radiation Oncology
. 2025 Oct 27;56:101063. doi: 10.1016/j.ctro.2025.101063

Deep learning [18F]-FDG-PET/CT‑based algorithm for tumor burden estimation in metastatic melanoma patients under immunotherapy

Lorenzo Lo Faro a,b,1, Hubert S Gabryś a,1, Simon Burgermeister a, Daniel Abler c,d, Maksym Fritsak a, Maiwand Ahmadsei a, Ciro Franzese b,e, Adrien Depeursinge d,f, Michel A Cuendet c,g,h, Stephanie Tanadini-Lang a, Panagiotis Balermpas a, Marta Scorsetti b, Matthias Guckenberger a, Sebastian M Christ a,
PMCID: PMC12639252  PMID: 41281625

Highlights

  • This study evaluates a deep learning model (PARS) for lesion detection and tumor burden estimation in metastatic melanoma.

  • PARS detected 68.9% of expert-identified lesions but with a modest precision of 46.8%.

  • Lung lesions showed highest precision (74.0%), while bone lesions had the lowest (32.9%).

  • Lesion volume estimates had good agreement (ICC = 0.77) but were generally underestimated.

  • Tumor burden estimation was highly variable and poorly correlated with expert assessments (ICC = 0.28).

Keywords: Deep-learning, Autosegmentation, Metastatic melanoma

Abstract

Background and purpose

Artificial intelligence is increasingly used in radiation oncology, yet its application for tumor burden (TB) estimation remains limited. This study evaluated the performance of a [18F]-fluorodeoxyglucose positron emission tomography/computerized tomography ([18F]-FDG-PET/CT)-based deep learning model, PET-Assisted Reporting System (“PARS”, Siemens Healthineers), for lesion detection, segmentation, and TB estimation in patients with metastatic melanoma undergoing immunotherapy.

Materials and methods

This retrospective study included 165 stage IV melanoma patients who underwent [18F]-FDG-PET/CT imaging prior to immunotherapy. Gross tumor volumes were segmented using PARS and compared with manual delineations performed by radiation oncologists. Performance was assessed through lesion detection metrics (precision and recall), individual lesion volume agreement, and overall TB estimation accuracy.

Results

PARS demonstrated an overall recall (sensitivity) of 68.9 %, though with modest precision (46.8 %). Performance was location-dependent, with highest precision observed for lung lesions (74.0 %) and lowest for bone lesions (32.9 %). For lesions detected by both methods, PARS tended to underestimate lesion volumes by an average (median) of 0.9 cc (median relative percentage difference (MRPD) =  −34.3 %), with a good agreement (intraclass correlations coefficient (ICC) = 0.77). The global TB in the whole cohort was overestimated by 28.3 %, but patient-level TB was on average (median) underestimated by 1.1 cc (MRPD =   −18.4 %) with high variability with a median absolute relative percentage difference (MARPD) = 68.6 %) and poor agreement (intraclass correlation coefficient (ICC) = 0.28).

Conclusions

PARS shows potential for treatment decision support with moderate accuracy in lesion detection and lesion volume estimation, but demonstrates significant variability in TB estimation, highlighting the need for further model refinements before clinical adoption.

Introduction

In recent years, artificial intelligence (AI) has emerged as a transformative force in radiation oncology, revolutionizing key areas such as the identification and contouring of organs at risk (OARs) and tumors [[1], [2], [3], [4], [5], [6], [7], [8], [9]], the development of radiotherapy treatment plans [[10], [11], [12]], and the assessment and prediction of treatment responses [[13], [14], [15], [16]]. Deep learning (DL) models, a subset of AI tools, have demonstrated the ability to streamline these processes, offering significant time savings and enhancing workflow efficiency across clinical settings.

Manual tumor segmentation, a critical step in radiotherapy planning and treatment monitoring, remains a time-consuming, labor-intensive task, that requires the expertise of highly trained clinicians. Studies evaluating automated segmentation have reported time savings ranging from 27 % in prostate cancer to as much as 93 % in head and neck cancers [17,18]. Despite these advances, tumor segmentation presents unique challenges compared to OAR segmentation. These challenges arise from tumor-specific factors, including size variability, e.g., in brain tumors, the intricate nature of surrounding anatomical structures, e.g., in head and neck tumors, and the frequent lack of clear tumor boundaries, e.g., in abdominal cancers.

The complexity of tumor segmentation further increases in oligo- and polymetastatic disease, where lesions appear in diverse anatomical locations and tumor environments and can be challenging to differentiate from neighboring regions of physiologically high glucose intake. In such settings, an effective automated segmentation tool must, once again, overcome significant heterogeneity in lesion size, shape, and surrounding tissues. A reliable and efficient model capable of accurately identifying individual lesion size changes, detecting new metastatic lesions, and quantifying total tumor burden (TB) could provide substantial benefits in radiation oncology throughout longitudinal patient monitoring, enabling rapid treatment adaptations, and ultimately improving patient outcomes [19].

Metastatic melanoma, the deadliest form of skin cancer, despite recent therapeutic advances, represents an ideal candidate for exploring AI-driven tumor segmentation and TB estimation. Mortality from melanoma is predominantly due to metastatic spread, which is typically assessed during follow-up via serial computer tomography (CT) or [18F]-fluorodeoxyglucose positron emission tomography [18F]-FDG-PET/CT imaging. Although automated tumor segmentation approaches have been studied in other malignancies, research in metastatic melanoma remains limited. Recently, Dirks et al. (2022) introduced a PET-based DL software for metastatic melanoma lesion detection and contouring, achieving promising results in segmentation accuracy and minimizing false positive and false negative rates [20].

This study aims to evaluate the performance of a novel PET-based DL software [[21], [22], [23], [24]], PET-Assisted Reporting System (“PARS”, Siemens Healthineers, [25,26]), for the detection and delineation of metastatic melanoma lesions. Additionally, it investigates the tool’s ability to provide precise TB estimations both at baseline and during follow-up imaging. To our knowledge, this study represents the largest cohort to date focusing on the objective quantification of TB in metastatic melanoma, contributing valuable insights to the field of AI applications in radiation oncology.

Materials and methods

Study design

Retrospective, single-center diagnostic accuracy study. The automated PET/CT segmentation model (PARS) was compared with expert manual segmentation as the reference standard. Primary outcome was lesion-level detection. The secondary outcome was agreement for lesion size and patient-level burden.

Patient cohort

This retrospective study included 165 patients diagnosed with stage IV melanoma who underwent treatment with either single (anti-PD-1; Nivolumab, or, anti-CTLA4; Ipilimumab) or dual checkpoint inhibition (anti-PD-1/anti-CTLA-4; Nivolumab/Ipilimumab) or a combination with an indoleamine 2-3-diosygenase (IDO) inhibitor (IDO1i; Epacadostat) between 2013 and 2019 at the Comprehensive Cancer Center Zurich (CCCZ). The study adhered to Good Clinical Practice (GCP) guidelines and the Declaration of Helsinki and was approved by the local ethics committee, the Kantonale Ethikkommission Zürich (KEK), approval number #2019-01012. Written informed consent was obtained from all participants.

Patient inclusion criteria were a confirmed metastatic melanoma diagnosis, presence of at least one non-brain lesion at the baseline, and availability of [18F]- FDG-PET/CT imaging. Patients were excluded if they lacked baseline or follow-up imaging, had only contrast-enhanced CT imaging, presented with brain metastases as the sole metastatic site, or had baseline lesions smaller than 0.5 cc. A detailed summary of patient demographics and clinical characteristics is provided in Table 1.

Table 1.

Patients’ characteristics.

All patients 165
Age (years)
Median 67
Q1–Q3 53–74
Range 28–93

Sex
Male 112 (67.9 %)
Female 53 (32.1 %)

Type of ICI
aPD1 120 (72.7 %)
aCTLA4 + aPD1 43 (26.1 %)
aCTLA4 1 (0.6 %)
aPD1 + ID01i 1 (0.6 %)

Number of lesions
Total 909
Median 3
Q1-Q3 1– 7
Range 1– 46

Metastatic sites
Lymph node 300 (33.0 %)
Lung 168 (18.5 %)
Liver 116 (12.8 %)
Bone 110 (12.1 %)
Other 215 (23.7 %)

Imaging protocols

All included patients underwent [18F]- FDG-PET/CT prior to treatment initiation. Imaging was conducted using standardized protocols across a single institution. [18F]-FDG doses were adjusted based on body weight or body mass index (BMI) (2.0–3.5 MBq per kg). The mean uptake time was 60 min.

The scans were acquired using a range of PET/CT scanners (GE Discovery MI (38 %), GE Discovery 690 (36 %), GE Discovery RX (12 %), GE Discovery STE (8 %), and other (6 %)). Pet images were reconstructed using OSEM variants with TOF and PSF when available in 63 % of studies. GE Q.Clear (BSREM) was used for the remaining 37 % of scans. In CT, no contrast was used. kVp was set to 120 kV and 140 kV in 78 % and 22 % of scans, respectively. Standard CT reconstruction kernel was used in 99 % of cases.

Lesion segmentation

Ground-truth segmentation was performed manually in three dimensions by two experienced radiation oncologists following a common protocol and relying on radiological reports. CT and PET images were rigidly registered, the PET signal was used for localization, CT for tissue contouring. CT-based contours were propagated onto PET images. Spatial mismatches were corrected manually, and PET-based contours were refined to conform to PET signal intensities.

Automated lesion segmentation was performed using PARS. PET regions with standard uptake value (SUV) peak values exceeding the mean blood pool uptake (SUVBP) by more than two standard deviations (SDs) were identified and segmented using a 42 % threshold of the local SUV max. Lesions smaller than 0.5 cc were excluded from further analysis. The convolutional neural network within PARS classified segmented foci as “benign” (model output < 0.5) or “suspicious” (model output ≥ 0.5). In this study we focused on analyzing, PARS segmentations marked as “suspicious” and did not evaluate the “benign” ones.

Statistical analysis

Manual segmentations served as the ground truth for evaluating the PARS model's performance. The PARS output included lesion segmentations and a probability score indicating the likelihood of each segmentation being “suspicious”. Each segmentation was classified as true positive or false positive based on comparison with expert annotations.

The accuracy of lesion detection was assessed using precision and recall (sensitivity). These performance metrics were computed for the entire cohort and further stratified by anatomical locations to assess region-specific performance variations. Additionally, precision and recall were evaluated across varying probability thresholds to illustrate the inherent trade-offs between sensitivity and specificity.

The accuracy of lesion detection count, individual lesion size estimation, and overall TB estimation by the PARS model were evaluated using several statistical measures: Median difference, median absolute difference (MAD), median relative percentage difference (MRPD), median absolute relative percentage difference (MARPD), and the intraclass correlation coefficient (ICC(3,1)) were calculated to quantify agreement and variability between PARS and expert measurements.

Results

Experts identified a total of 955 lesions across 165 patients (Table 1). The PARS model identified 6456 benign and 1377 suspicious lesions, with the classification threshold set at the default value of 0.5 to distinguish between benign and suspicious lesions.

In several cases, multiple small segmentations by PARS corresponded to a single larger segmentation by the experts, or vice versa. To establish a consistent one-to-one correspondence between expert and PARS segmentations, smaller segmentations were merged. Following this process, the number of expert-defined lesions was reduced to 909, and the unique PARS segmentations were reduced to 6450 benign and 1337 suspicious lesions.

Overall, PARS identified 1337 suspicious lesions, representing a 47.1 % overestimation compared to the experts' count. The model's overall precision was 46.8 %, indicating that more than half of the lesions detected by PARS were false positives (n = 711). The overall recall (sensitivity) was 68.9 %, as the model successfully detected 626 of the 909 lesions identified by experts (Table 2). Visual comparison of expert and PARS-based segmentations of a few selected cases is presented in Fig. 1.

Table 2.

Precision and recall of the PARS model in detecting lesions overall and stratified by anatomical location. Precision represents the proportion of correctly identified lesions among all lesions identified by PARS, while recall indicates the proportion of true lesions identified by PARS. Bold values indicate the highest recall and precision among the evaluated anatomical sites.

Location Recall Precision
All 68.9 % 46.8 %
Lymph node 71.0 % 47.8 %
Bone 74.5 % 32.9 %
Liver 81.9 % 43.8 %
Lung 57.7 % 74.0 %

Fig. 1.

Fig. 1

Examples of the differences between expert and PARS segmentations. A: Substantial underestimation of TB by PARS; B: Good match between experts and PARS; C: Substantial number of false positives and overestimation of TB.

Lesion detection performance varied significantly depending on anatomical location (Table 2). The highest precision was observed for lung lesions (74.0 %), indicating effective identification with fewer false positives. In contrast, bone lesions had the lowest precision (32.9 %), reflecting a substantial number of false positives. Liver and lymph node lesions showed intermediate precision values of 43.8 % and 47.8 %, respectively. Despite its low precision, the model demonstrated relatively high recall for bone lesions (74.5 %), indicating good sensitivity, but difficulty in distinguishing true lesions from false positives.

Performance trade-offs between precision and recall at various lesion classification probability thresholds were also analyzed (Fig. 2). Increasing the threshold improved precision by reducing false positives, but simultaneously decreased recall, leading to fewer true lesions being identified. This trend highlights the inherent challenge in balancing sensitivity and specificity in automated lesion detection.

Fig. 2.

Fig. 2

Precision and recall of PARS model. A: Precision and recall of the PARS model plotted against the PARS probability threshold for different tumor sites. B: Precision–recall curves for each site, illustrating the relationship between precision and recall across all possible probability thresholds.

At the patient level, PARS overestimated the lesion count by an average of approximately one lesion per patient, albeit with substantial variability (Table 3, Fig. 3). The ICC for lesion counts per patient was 0.68, indicating moderate agreement between PARS and expert assessments.

Table 3.

Agreement between the PARS model and expert annotations for lesion detection, individual lesion size estimation, and total tumor burden estimation. Metrics include median difference, median absolute difference (MAD), median relative percentage difference (MRPD), median absolute relative percentage difference (MARPD), and the intraclass correlation coefficient (ICC(3,1)) with 95% confidence intervals (CI).

Median difference Median absolute difference (MAD) Median relative percentage difference (MRPD) Median absolute relative percentage difference (MARPD) ICC (95 % CI)
Lesion detection 1 2 12.5 % 57.1 % 0.68 (0.60– 0.76)
Lesion size [cc] −0.9 1.5 −34.3 % 44.9 % 0.77 (0.55–0.88)
Tumor burden [cc] −1.1 14.0 −18.4 % 68.6 % 0.28 (0.12–0.51)

Fig. 3.

Fig. 3

Comparison of PARS and expert assessments for three metrics: (A, B, C) the number of lesions, (D, E, F) the lesion size (in cc), and (G, H, I) the total tumor burden (in cc). The top row of boxplots (A, D, G) shows the absolute difference (PARS – Experts), while the middle row (B, E, H) shows the relative percentage error on a symlog scale. In each boxplot, the central line denotes the median, box edges indicate the interquartile range, and whiskers extend to 1.5 × IQR. Outliers are shown as open circles. The bottom row (C, F, I) displays scatter plots of PARS versus Experts, with the diagonal line indicating equality.

For lesions detected by both PARS and experts (n = 626; 68.9 % of expert-identified lesions), PARS tended to underestimate lesion volume by an average of 0.9 cc (MRPD =  −34.3 %; Table 3). The ICC of 0.77 suggests relatively good agreement for individual lesion size estimates.

Stratified analysis by lesion location (Table S1), showed that the median absolute error was greatest for liver lesions (3.9 cc) and smallest for lung lesions (0.7 cc). Percentage-based analyses revealed low systematic bias, but high variability for liver lesions (MRPD =  −6.1 %, MARPD =  56.4 %), whereas lung lesions were consistently undersegmented with lower variability (MRPD =  −22.0 %, MARPD =  38.0 %). The highest ICC, indicating excellent agreement, was achieved for liver lesions (ICC =  0.90), whereas lung lesions had the lowest ICC =  0.52.

Across the entire cohort, PARS overestimated the total TB by 28.3 %. However, at the individual patient level, the model generally underestimated TB by an average of 1.1 cc (MRPD =  −18.4 %; Table 3, Fig. 3). The discrepancy between overall and individual-level findings results from high variability in estimates between patients. This variability was also reflected in a high median absolute difference (MAD) = 14.0 % and mean absolute relative percentage error (MARPE) = 68.6 %. The ICC for TB estimation was low (0.28), indicating poor overall agreement between PARS and expert assessments.

Discussion

This study evaluated the performance of the PARS model for detecting, segmenting, and estimating TB in metastatic melanoma, focusing on its application in a real-world cohort. PARS achieved a recall of 68.9 %, identifying most true lesions, but its precision was limited (46.8 %) due to high FP rates. Anatomical site-specific performance varied, with lung lesions exhibiting the highest precision (74.0 %) and bone lesions the lowest (32.9 %), despite high recall (74.5 %) in the latter. Overall, the agreement between the PARS model and expert annotations demonstrated moderate to good performance in lesion detection (ICC =  0.68) and individual lesion size estimation (ICC =  0.77). However, the agreement for estimating total TB was considerably lower (ICC =  0.28), indicating limited reliability of the model in accurately assessing overall TB.

The high FP rate highlights a common challenge in DL-based segmentation, differentiating between pathological and physiological [18F]-FDG uptake [[4], [5], [6]]. Bone lesions, which exhibited particularly low precision, likely reflect the inherent complexity of bone imaging due to overlapping signals from bone marrow activity and metabolic processes. Similarly, lymph node segmentation may suffer from challenges in distinguishing benign hyperplasia from malignant involvement. However, it is important to bear in mind that only lymph node lesions > 0.7 cm were manually contoured by physicians. Both recall (68.9 %) and precision (46.8 %) in this study compare unfavorably to previous studies. For example, according to Sibille et al. (2020), PARS achieved an area under the curve (AUC) of the receiver-operating characteristic of >0.95 for lesion classification and anatomic localization, and accuracies of 96 % for body part and 87 % for organ localization, respectively, in lymphoma and lung cancer cases, for which PARS was originally trained. Addressing some of these issues might involve refining the classification algorithm within PARS to incorporate multi-parametric features, such as lesion shape, heterogeneity, and proximity to critical anatomical landmarks, which could help improve specificity without sacrificing recall.

With respect to TB estimation, a similar observation can be made: Capobianco et al. (2021) found that TB estimation in lymphoma patients, for which the model was originally built, in this particular case with Diffuse Large B-Cell Lymphoma (DLBCL), was consistent with that obtained by experts and displayed significant prognostic value for both progression-free and overall survival [24]. In contrast, Abler et al. (2023) reported comparable results to ours in terms of detection precision (53 %) and negative-predictive-value (80 %) when assessing malignant melanoma lesions, even though the imaging data set was less granular (not by organ) and based on a review of a small set of automatic segmentations (24 images) [27]. Of note, the authors also found that detection accuracy improved when combining predictions at different time points. While PARS has not been trained extensively on metastatic melanoma patients, the variability in performance across anatomical sites nonetheless points to limitations in PARS' current approach to volumetric analysis. Smaller lesions and lesions with irregular boundaries, which were more prone to overestimation, may benefit from improved contour refinement algorithms. Incorporating edge-detection techniques and probabilistic models to delineate unclear boundaries could enhance segmentation accuracy [28]. Furthermore, using multi-modal imaging inputs—such as combining PET with MRI to provide more detailed soft-tissue contrast—could improve delineation in regions like the liver and lymph nodes [29].

Moreover, future iterations of PARS could leverage newer DL architectures such as transformer models, which have shown promise in other medical imaging applications [30]. Transformers are capable of modeling long-range dependencies, making them well-suited for detecting subtle differences in signal intensity and spatial relationships between lesions and surrounding tissues. Additionally, incorporating unsupervised or semi-supervised learning approaches could help overcome the scarcity of annotated data, particularly for rare or atypical lesion presentations [31].

Last, but not least, active learning, where the model queries clinicians to annotate uncertain cases during training, could also be implemented to iteratively improve PARS segmentation performance [32]. This approach would help the model learn from edge cases, such as atypical metastatic patterns or subtle disease presentations, which are often the most challenging to detect and rightly classify.

The utility of PARS extends beyond lesion detection and segmentation. In clinical settings, rapid and accurate TB estimation could play a pivotal role in treatment planning and monitoring, particularly for checkpoint inhibitor therapies where early response assessment is critical. While PARS currently requires significant expert oversight, future versions could enable semi-automated workflows where the software generates initial segmentations for clinician review, thereby reducing the manual workload without compromising accuracy.

The precision-recall trade-off observed in this study suggests that dynamic thresholding might be implemented to tailor segmentation outputs to specific clinical needs. For instance, lower thresholds could be used for comprehensive disease screening, while higher thresholds might guide more targeted applications such as stereotactic body radiotherapy planning.

This retrospective, single-center study has several limitations. PET/CT scans were acquired on multiple scanners without phantom harmonization was performed. Manual contours from two radiation oncologists served as ground truth, but interobserver agreement was not quantified. The cohort excluded patients with brain-only metastases and came from a single center, which limits generalizability.

In summary, to fully realize its potential, PARS and similar AI tools must address the limitations observed in this study: (1) Enhanced training datasets: Expanding the training set to include diverse lesion types, sizes, and anatomical contexts could improve model generalizability. Annotated data from multi-institutional cohorts, particularly for underrepresented lesion sites, would be invaluable; (2) Expansion of scope: Using information from the available CT scans for lesion contouring might be the next pragmatic step forward; (3) Multi-modal imaging: Integrating data from different imaging modalities, such as MRI or contrast-enhanced CT, could enhance lesion characterization and reduce FPs; (4) Adaptive learning: Models that adapt to patient-specific characteristics, such as lesion growth patterns and metabolic activity changes, could improve segmentation accuracy and TB estimation over time; and (4) Clinical decision support systems: Future iterations of PARS could incorporate predictive analytics, such as estimating the likelihood of treatment response or disease progression based on TB dynamics, to assist medical, radiation and surgical oncologists in treatment planning.

In conclusion, while the PARS model demonstrates promising capabilities for lesion detection and TB estimation, it requires further refinement to overcome site-specific limitations and variability in performance. Incorporating advanced AI techniques, diverse training datasets, and multi-modal imaging could substantially enhance its utility. With continued development, tools like PARS have the potential to change how clinicians monitor and manage metastatic melanoma, ultimately improving patient outcomes.

Availability of data and material

Collected patient and imaging data are confidential and not available for publication.

Code availability

Not applicable for this publication.

Authors' contributions

All authors made significant contributions to this project. The idea and conceptualization of the project were developed by STL, MG and SMC. STL and MG obtained the ethical approval for the project. HSG, MA, SMC, SB, and others supported the manual lesion segmentation. LLF, HSG, and MF applied the PARS model to the imaging data set and conducted all subsequent data analysis and visualization efforts. LLF, HSG and SMC prepared the manuscript. The manuscript was then critically reviewed by all co-authors, incl. CF, PB, and MS. The final version of the manuscript was approved by all authors before submission to PhiRO.

Funding

Lorenzo Lo Faro received a fellowship grant from the Department of Radiation Oncology, University of Zurich, to enable him to work on this project.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgement

The project was also supported through the C3Z Precision Oncology Funding Program as part of “The CCCZ oligometastatic disease program – OMDZH”.

Presentations

The abstract of this manuscript was submitted for the ESTRO 2025 Annual Congress in Vienna, Austria.

Ethics approval

This study was approved by the Swiss Cantonal Ethics Committee before the initiation of the project (BASEC ID# 2019-01012).

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.ctro.2025.101063.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Supplementary Data 1
mmc1.docx (1.9MB, docx)

References

  • 1.Zhou Z., Jain P., Lu Y., Macapinlac H., Wang M.L., Son J.B., et al. Computer-aided detection of mantle cell lymphoma on (18)F-FDG PET/CT using a deep learning convolutional neural network. Am J Nucl Med Mol Imaging. 2021;11(4):260–270. [PMC free article] [PubMed] [Google Scholar]
  • 2.Liu C., Gardner S.J., Wen N., Elshaikh M.A., Siddiqui F., Movsas B., et al. Automatic segmentation of the prostate on CT images using deep neural networks (DNN) Int J Radiat Oncol Biol Phys. 2019 Jul;104(4):924–932. doi: 10.1016/j.ijrobp.2019.03.017. [DOI] [PubMed] [Google Scholar]
  • 3.Liu X., Li K.-W., Yang R., Geng L.-S. Review of deep learning based automatic segmentation for lung cancer radiotherapy. Front Oncol. 2021;11 doi: 10.3389/fonc.2021.717039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Vrtovec T., Močnik D., Strojan P., Pernuš F., Ibragimov B. Auto-segmentation of organs at risk for head and neck radiotherapy planning: from atlas-based to deep learning methods. Med Phys. 2020 Sep;47(9):e929–e950. doi: 10.1002/mp.14320. [DOI] [PubMed] [Google Scholar]
  • 5.Schouten JPE, Noteboom S, Martens RM, Mes SW, Leemans CR, de Graaf P, et al. Automatic segmentation of head and neck primary tumors on MRI using a multi-view CNN. Cancer Imaging [Internet]. 2022;22(1):8. Available from: https://doi.org/10.1186/s40644-022-00445-7. [DOI] [PMC free article] [PubMed]
  • 6.Rodríguez Outeiral R., Bos P., Al-Mamgani A., Jasperse B., Simões R., van der Heide U.A. Oropharyngeal primary tumor segmentation for radiotherapy planning on magnetic resonance imaging using deep learning. Phys Imaging Radiat Oncol. 2021 Jul;19:39–44. doi: 10.1016/j.phro.2021.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Xue J., Wang B., Ming Y., Liu X., Jiang Z., Wang C., et al. Deep learning-based detection and segmentation-assisted management of brain metastases. Neuro Oncol. 2020 Apr;22(4):505–514. doi: 10.1093/neuonc/noz234. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Yue W., Zhang H., Zhou J., Li G., Tang Z., Sun Z., et al. Deep learning-based automatic segmentation for size and volumetric measurement of breast cancer on magnetic resonance imaging. Front Oncol. 2022;12 doi: 10.3389/fonc.2022.984626. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Yue Y., Li N., Zhang G., Zhu Z., Liu X., Song S., et al. Automatic segmentation of esophageal gross tumor volume in (18)F-FDG PET/CT images via GloD-LoATUNet. Comput Methods Programs Biomed. 2023 Feb;229 doi: 10.1016/j.cmpb.2022.107266. [DOI] [PubMed] [Google Scholar]
  • 10.Wang M., Zhang Q., Lam S., Cai J., Yang R. A review on application of deep learning algorithms in external beam radiotherapy automated treatment planning. Front Oncol. 2020;10 doi: 10.3389/fonc.2020.580919. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Johnston N., De Rycke J., Lievens Y., van Eijkeren M., Aelterman J., Vandersmissen E., et al. Dose-volume-based evaluation of convolutional neural network-based auto-segmentation of thoracic organs at risk. Phys Imaging Radiat Oncol [internet] 2022;23:109–117. doi: 10.1016/j.phro.2022.07.004. https://www.sciencedirect.com/science/article/pii/S2405631622000677 Available from. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Zhang S., Lv B., Zheng X., Li Y., Ge W., Zhang L., et al. Dosimetric study of deep learning-guided ITV prediction in cone-beam CT for lung stereotactic body radiotherapy. Front Public Heal. 2022;10 doi: 10.3389/fpubh.2022.860135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Peeken J.C., Asadpour R., Specht K., Chen E.Y., Klymenko O., Akinkuoroye V., et al. MRI-based delta-radiomics predicts pathologic complete response in high-grade soft-tissue sarcoma patients treated with neoadjuvant therapy. Radiother Oncol J Eur Soc Ther Radiol Oncol. 2021 Nov;164:73–82. doi: 10.1016/j.radonc.2021.08.023. [DOI] [PubMed] [Google Scholar]
  • 14.Cho J., Kim Y.J., Sunwoo L., Lee G.P., Nguyen T.Q., Cho S.J., et al. Deep learning-based computer-aided detection system for automated treatment response assessment of brain metastases on 3D MRI. Front Oncol. 2021;11 doi: 10.3389/fonc.2021.739639. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Pang X., Wang F., Zhang Q., Li Y., Huang R., Yin X., et al. A pipeline for predicting the treatment response of neoadjuvant chemoradiotherapy for locally advanced rectal cancer using single MRI modality: combining deep segmentation network and radiomics analysis based on “Suspicious Region”. Front Oncol. 2021;11 doi: 10.3389/fonc.2021.711747. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Jayachandran Preetha C., Meredig H., Brugnara G., Mahmutoglu M.A., Foltyn M., Isensee F., et al. Deep-learning-based synthesis of post-contrast T1-weighted MRI for tumour response assessment in neuro-oncology: a multicentre, retrospective cohort study. Lancet Digit Heal. 2021 Dec;3(12):e784–e794. doi: 10.1016/S2589-7500(21)00205-3. [DOI] [PubMed] [Google Scholar]
  • 17.Martin S., Rodrigues G., Patil N., Bauman G., D’Souza D., Sexton T., et al. A multiphase validation of atlas-based automatic and semiautomatic segmentation strategies for prostate MRI. Int J Radiat Oncol Biol Phys. 2013 Jan;85(1):95–100. doi: 10.1016/j.ijrobp.2011.07.046. [DOI] [PubMed] [Google Scholar]
  • 18.Oktay O, Nanavati J, Schwaighofer A, Carter D, Bristow M, Tanno R, et al. Evaluation of deep learning to augment image-guided radiotherapy for head and neck and prostate cancers. JAMA Netw Open [Internet]. 2020 Nov 30;3(11):e2027426–e2027426. Available from: https://doi.org/10.1001/jamanetworkopen.2020.27426. [DOI] [PMC free article] [PubMed]
  • 19.Kim J, Chang JS, Sung W, Kim JS, Kim TH, Choi SH, et al. A Comparison of 2 disease burden assessment methods (3D volume vs the number of lesions) for prognostication of survival in metastatic melanoma: implications for the characterization of oligometastatic disease. Int J Radiat Oncol Biol Phys [Internet]. 2022;V. Available from: https://doi.org/10.1016/j.ijrobp.2022.08.040. [DOI] [PubMed]
  • 20.Dirks I., Keyaerts M., Neyns B., Vandemeulebroucke J. Computer-aided detection and segmentation of malignant melanoma lesions on whole-body (18)F-FDG PET/CT using an interpretable deep learning approach. Comput Methods Programs Biomed. 2022 Jun;221 doi: 10.1016/j.cmpb.2022.106902. [DOI] [PubMed] [Google Scholar]
  • 21.Sibille L., Seifert R., Avramovic N., Vehren T., Spottiswoode B., Zuehlsdorff S., et al. (18)F-FDG PET/CT uptake classification in lymphoma and lung cancer by using deep convolutional neural networks. Radiology. 2020 Feb;294(2):445–452. doi: 10.1148/radiol.2019191114. [DOI] [PubMed] [Google Scholar]
  • 22.Pinochet P., Eude F., Becker S., Shah V., Sibille L., Toledano M.N., et al. Evaluation of an automatic classification algorithm using convolutional neural networks in oncological positron emission tomography. Front Med. 2021;8 doi: 10.3389/fmed.2021.628179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Weber M., Kersting D., Umutlu L., Schäfers M., Rischpler C., Fendler W.P., et al. Just another “Clever Hans”? Neural networks and FDG PET-CT to predict the outcome of patients with breast cancer. Eur J Nucl Med Mol Imaging. 2021 Sep;48(10):3141–3150. doi: 10.1007/s00259-021-05270-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Capobianco N., Meignan M., Cottereau A.-S., Vercellino L., Sibille L., Spottiswoode B., et al. Deep-learning (18)F-FDG uptake classification enables total metabolic tumor volume estimation in diffuse large B-cell lymphoma. J Nucl Med. 2021 Jan;62(1):30–36. doi: 10.2967/jnumed.120.242412. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Siemens Healthineers. How we developed Auto ID [Internet]. 2021. Available from: https://cdn0.scrvt.com/39b415fb07de4d9656c7b516d8e2d907/e5b3e064a69289f2/7c4665ca257c/siemens-healthineers_mi_auto_id_artificial_intelligence_petct.pdf#:∼:text=certified radiologist from the University, on the future of medicine.
  • 26.Siemens Healthineers. Lesion Scout with Auto ID in syngo.via MM Oncology [Internet]. 2021. Available from: https://marketing.webassets.siemens-healthineers.com/1e0fb011cca5bd1d/835a53b2c03b/siemens-healthineers_mi_syngo_via_lesion_scout_with_auto_id_mm_oncology_white_paper_feb-21.pdf.
  • 27.Abler D., Courlet P., Dietz M., Gatta R., Girard P., Munafo A., et al. Semiautomated pipeline to quantify tumor evolution from real-world positron emission tomography/computed tomography imaging. JCO Clin Cancer Informatics. 2023 May;7 doi: 10.1200/CCI.22.00126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Rayed M.E., Islam S.M.S., Niha S.I., Jim J.R., Kabir M.M., Mridha M.F. Deep learning for medical image segmentation: state-of-the-art advancements and challenges. Informatics Med Unlocked [internet] 2024;47 https://www.sciencedirect.com/science/article/pii/S2352914824000601 Available from. [Google Scholar]
  • 29.Ren J., Eriksen J.G., Nijkamp J., Korreman S.S. Comparing different CT, PET and MRI multi-modality image combinations for deep learning-based head and neck tumor segmentation. Acta Oncol. 2021 Nov;60(11):1399–1406. doi: 10.1080/0284186X.2021.1949034. [DOI] [PubMed] [Google Scholar]
  • 30.Xiao H., Li L., Liu Q., Zhu X., Zhang Q. Transformers in medical image segmentation: a review. Biomed Signal Process Control [internet] 2023;84 https://www.sciencedirect.com/science/article/pii/S1746809423002240 Available from. [Google Scholar]
  • 31.He K., Gan C., Li Z., Rekik I., Yin Z., Ji W., et al. Transformers in medical image analysis. Intell Med [internet] 2023;3(1):59–78. https://www.sciencedirect.com/science/article/pii/S2667102622000717 Available from. [Google Scholar]
  • 32.Wang H., Jin Q., Li S., Liu S., Wang M., Song Z. A comprehensive survey on deep active learning in medical image analysis. Med Image Anal [internet] 2024;95 doi: 10.1016/j.media.2024.103201. https://www.sciencedirect.com/science/article/pii/S1361841524001269 Available from. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Data 1
mmc1.docx (1.9MB, docx)

Data Availability Statement

Collected patient and imaging data are confidential and not available for publication.


Articles from Clinical and Translational Radiation Oncology are provided here courtesy of Elsevier

RESOURCES