Skip to main content
Cambridge Prisms: Precision Medicine logoLink to Cambridge Prisms: Precision Medicine
. 2022 Nov 11;1:e4. doi: 10.1017/pcm.2022.9

Artificial intelligence for early detection of renal cancer in computed tomography: A review

William C McGough 1,2,, Lorena E Sanchez 3,4, Cathal McCague 1,3,4, Grant D Stewart 4,5, Carola-Bibiane Schönlieb 6, Evis Sala 3,4, Mireia Crispin-Ortuzar 1,2
PMCID: PMC10953744  PMID: 38550952

graphic file with name S2752614322000096_figAb.jpg

Keywords: computed tomography, deep learning, machine learning, medical imaging, cancer

Abstract

Renal cancer is responsible for over 100,000 yearly deaths and is principally discovered in computed tomography (CT) scans of the abdomen. CT screening would likely increase the rate of early renal cancer detection, and improve general survival rates, but it is expected to have a prohibitively high financial cost. Given recent advances in artificial intelligence (AI), it may be possible to reduce the cost of CT analysis and enable CT screening by automating the radiological tasks that constitute the early renal cancer detection pipeline. This review seeks to facilitate further interdisciplinary research in early renal cancer detection by summarising our current knowledge across AI, radiology, and oncology and suggesting useful directions for future novel work. Initially, this review discusses existing approaches in automated renal cancer diagnosis, and methods across broader AI research, to summarise the existing state of AI cancer analysis. Then, this review matches these methods to the unique constraints of early renal cancer detection and proposes promising directions for future research that may enable AI-based early renal cancer detection via CT screening. The primary targets of this review are clinicians with an interest in AI and data scientists with an interest in the early detection of cancer.

Impact statement

Initially, this review discusses existing approaches in automated renal cancer diagnosis, and methods across broader AI research, to summarise the existing state of AI cancer analysis. Then, this review matches these methods to the unique constraints of early renal cancer detection and proposes promising directions for future research that may enable AI-based early renal cancer detection via CT screening.

Introduction

In 2017, 393,000 renal cancer (RC) diagnoses and 139,000 RC deaths were recorded worldwide (Fitzmaurice et al., 2019). Renal cell carcinoma (RCC), the most common cancer involving the kidney, is mostly discovered incidentally during routine health checks or in the assessment of unrelated symptoms, and patients with incidentally discovered RCC tend to have better health outcomes than those diagnosed with symptomatic RCC (Rabjerg et al., 2014; Vasudev et al., 2020). This is because symptom presentation is generally associated with later-stage progression (Rabjerg et al., 2014; Vasudev et al., 2020). As shown in Table 1, RC screening satisfies many of the 10 Wilson–Junger criteria of an effective screening program (Wilson et al., 1968; Rossi et al., 2018); in principle, regular RC screening could improve general survival rates by increasing the rate of early RC discovery.

Table 1.

The current state of satisfaction of Wilson–Junger criteria for AI RC screening in LDCT

Wilson–Junger criterion Satisfied by AI-based RC screening in LDCT (Y/?/N) Justification
1. Tackles an important health problem Y Roughly 139,000 RC deaths in 2017 (Fitzmaurice et al., 2019)
2. Can be followed by an acceptable treatment Y Treatment for RC is long-accepted and guideline-based (Ljungberg et al., 2015; Guidelines for the Management of Renal Cancer, 2016)
3. Facilities for diagnosis and treatment are available Y Patients treated with earlier diagnosis have higher survival rates; malignancy-diagnosing needle biopsies are highly accurate (Rabjerg et al., 2014)
4. A recognisable early stage exists Y 55–60% of all RC diagnoses are incidental (Rabjerg et al., 2014; Vasudev et al., 2020); RC initially exists in slow-growing latent state for many patients (Zhang et al., 2016)
5. Suitable test or examination are available N Development of AI technology for LDCT screening is required
6. Test is acceptable to population ? LDCT (among other screening methods) appears to be acceptable in screening the general population (Harvey-Kelly et al., 2020; Freer-Smith et al., 2021)
7. Natural history of disease is well understood N The natural history of early RC is only partially understood (Volpe et al., 2004; Zhang et al., 2016; Rossi et al., 2018)
8. Agreed policy on treatment Y Treatment policies for RC are well established (Ljungberg et al., 2015; Guidelines for the Management of Renal Cancer, 2016)
9. Acceptable cost-effectiveness ? AI technology is not yet developed; high costs are likely in manual CT screening (i.e. without AI) (Beinfeld et al., 2005)
10. Test is suitable as a continuing process ? NLST reduced participant mortality over 7-years of continuous screening; no equivalent study has been performed for RC (NLST, 2011)

Note. Y: Yes, currently satisfied; ?: Unknown, more research is needed to clarify; N: No, currently unsatisfied.

However, there are significant challenges associated with deploying the current standard method for RC discovery, contrast-enhanced computed tomography (CECT; Ljungberg et al., 2015; Guidelines for the Management of Renal Cancer, 2016), in RC screening: the high cost of computed tomography (CT) screening (Beinfeld et al., 2005; Ishikawa et al., 2007; Jensen et al., 2020), the risks of routine radiation exposure (Hunink and Gazelle, 2003), the lack of a definite target screening population (Rossi et al., 2018), and the low incidence of RC in the general population (O’Connor et al., 2011, 2018). These facts undermine LDCT’s cost-effectiveness and suitability for ongoing screening – Wilson–Junger criteria 9 and 10, respectively. Nevertheless, recent literature has indicated that cancer screening with low-dose computed tomography (LDCT) may improve population health and studies are ongoing in this area (NLST, 2011; Black et al., 2014; Stewart, 2021). Furthermore, developments in artificial intelligence (AI) have enabled the automation of some radiological tasks that may reduce the cost of CT analysis. Following these developments, this manuscript reviews AI technologies across automated RC diagnosis, other cancer domains, and broader computer vision to suggest novel research directions that may enable RC early detection in LDCT and non-contrast CT (NCCT), by automating and reducing the cost of analyses inherent to CT screening.

In this review, we define ‘early detection’ as the processes requisite in screening that detect early signs of disease in asymptomatic individuals. Image-based early detection and diagnosis may share many sub-processes, such as pre-processing, segmentation, radiomic feature extraction, post-processing, and classification. Within these sub-processes, segmentation and classification are the subjects of most machine learning research. Segmentation algorithms receive images as input and assign to them element-wise labels according to predefined semantic values, providing structure to images by highlighting the most salient regions of interest (ROI), making automated analyses simpler. An example of two-dimensional segmentation is shown in Figure 1. Classification refers to any process that assigns a discrete category to a data source; classification algorithms receive quantitative data (e.g., radiomic features, morphological measurements from a histology slide, or raw pixel data from an image) and assign a label to the data source; this label can be binary (malignant/benign) or multi-class (differentiating between RCC subtypes).

Figure 1.

Figure 1.

A segmented CECT axial slice, depicting the segmented kidneys (blue) and tumour(red). CT data taken from KiTS19, case 49.

Early detection methods must be cheap to be viable in screening. They must also be accurate, to detect a high rate of the target disease whilst minimising the rate of overdiagnosis, which can dramatically increase screening costs. AI analyses are automated by default, making them cheap enough to be operationally viable in screening. Therefore, the development of an AI-based RC early detection system should focus on optimising the AI system’s accuracy to maximise the system’s utility in screening.

This manuscript reviews existing AI diagnostic methods that may be suitable for early detection, and suggests possible improvements to these existing methods, due to the lack of existing AI research in RC early detection. The literature reviewed in this manuscript was extracted from three different sources, namely (i) Kidney and Tumour Segmentation Challenge (KiTS) winning submissions; (ii) ImageNet (March 2022), including four contemporary, high-scoring algorithms and four other highly cited algorithms often used in medical AI, and (iii) renal segmentation and classification articles (Google Scholar, January 2015–March 2022). A list of all papers initially selected for reading, and then finally included, in this review can be found in the Supplementary Material. The review is complemented by highly cited articles from other early detection domains, that may represent novel approaches for conducting AI LDCT screening for RC, and the broader AI literature, including hyperparameter optimisation, multi-task learning (MTL), and synthetic image generation.

AI primer

AI refers to any computational, data-driven decision-making system that enables the automation of complex tasks – mimicking human intelligence – without explicit instruction. Machine-learning models are a subset of AI systems that automatically learn to structure and/or make predictions, or ‘inferences’, from data. Supervised learning models learn using labelled datasets – a set of paired inputs and labelled outputs. In segmentation, labelled datasets contain CT scans and volumes of corresponding voxel-wise labels for each scan. Supervised learning models review labelled data during ‘training’, iteratively assessing each sample and altering its own mathematical parameters to progressively improve inference accuracy. Following training, a supervised learning model’s accuracy is evaluated over an unseen ‘validation’ labelled dataset, where the differences between the model’s inferences and the dataset’s labels are evaluated to determine the model’s overall accuracy. This manuscript exclusively reviews supervised machine-learning methods but, for brevity, ‘AI’ will be used as a general term for all models.

In classification and segmentation, the model’s responses can be categorised as true positive (TP), true negative (TN), false positive (FP), or false negative (FN). Accuracy metrics are derived from the ratio of these response classifications, such as sensitivity and specificity,

graphic file with name S2752614322000096_eqn1.jpg (1)
graphic file with name S2752614322000096_eqn2.jpg (2)

where Inline graphic refers to the number of Inline graphic observed in validation. Optimum performance usually requires a trade-off between maximum specificity and maximum sensitivity; the area under the receiver operating characteristic curve (AUC) and the Dice similarity coefficient (DSC) are commonly used accuracy metrics that quantify the model’s trade-off between specificity and sensitivity. AUC is generated by plotting the model’s receiver operating characteristic (specificity vs. sensitivity) and calculating the area under its curve; an example ROC is shown in Figure 2 for the reader’s understanding. Segmentation performance is generally evaluated by the DSC metric, defined by

graphic file with name S2752614322000096_eqn3.jpg (3)

Contemporary AI algorithms in image analysis tend to be comprised of convolutional neural networks (CNN) and/or transformers. This manuscript will not discuss the technical differences between these models, beyond the functional differences that exist with respect to their typical performance and cost characteristics. Both are deep learning algorithms (DL), meaning they are both types of neural network. The cost of CNNs scales linearly with the number of input image elements, whereas transformer cost scales quadratically, making transformers-only models much costlier during analyses of 3D images, such as in CT. Transformers can achieve ‘global’ attention and detect patterns across whole input images simultaneously, whereas CNNs can only achieve ‘local’ pattern recognition, as they must divide input images into smaller sections and analyse them individually. This leads to superior image analysis in transformer models where patterns have global interdependencies. The performance of CNN- and transformer-based models will be reviewed in this manuscript, as well as hybrid models that attempt to combine the benefit of both approaches.

Figure 2.

Figure 2.

An example ROC curve for an arbitrary classifier, displaying the trade-off between sensitivity and specificity in an arbitrary classification task. The further the curve is from the x-axis, and the closer it is to the y-axis, the higher the classifier’s holistic accuracy and AUC. In the shown ROC curve, AUC is 0.699.

AI in renal cell carcinoma diagnosis

Segmentation

Renal segmentation has received increased research attention following the advent of KiTS, first established in 2019 (Heller et al., 2019) and renewed in 2021. KiT19 and KiTS21 publicly released 210 CECT volumes and 300 CECT volumes, respectively, where all CT scans contained tumours and some contained cysts, and invited participants to submit their renal segmentation algorithms to compete in a fair assessment of accuracy. KiTS19’s winner, based on nnU-Net (Isensee and Maier-Hein, 2019; Isensee et al., 2021), was derived from the state-of-the-art segmentation CNN U-Net (Ronneberger et al., 2015) and focused on the optimisation of its hyperparameters (properties relating to training and model size) without altering the essential structure of U-Net. This approach represented a breakaway from the hitherto standard across segmentation research, of proposing modular architectural changes to U-Net for marginal accuracy gains. Outside of KiTS, nnU-Net scored highly in a wide variety of segmentation domains, winning other medical segmentation competitions across multiple organ sites (Isensee et al., 2021), proving the primacy of hyperparameter optimisation in maximising segmentation performance.

All KiTS21’s top-7 performing submissions made direct use of nnU-Net as a baseline algorithm. The top-3 submissions used nnU-Net’s ‘course-to-fine’ cascade approach. In this approach, a ‘course’ U-Net segments the input CT images at a low resolution to dictate an initial ROI; then, this segmentation inference is refined at higher resolutions by more ‘fine’ U-Nets. This process is repeated until the ROIs are labelled at full resolution. Figure 3 shows the performance distributions of KiTS19 and KiTS21’s top-7 submissions in renal segmentation – the adoption of nnU-Net significantly increased the mean mass-segmentation DSC among top performers (p = 3.58 × 10−5) from 0.832 in KiTS19 to 0.870 in KiTS21 (Challenge Leaderboard, 2019; KiTS21, 2021). To the authors’ knowledge, no other kidney segmentation algorithm has significantly improved upon KiTS21’s competition-winning nnU-Net-based approach (Zhao et al., 2022).

Figure 3.

Figure 3.

The performance distribution of the top-7 algorithms in KiTS19 and KiTS21, with respect to mass segmentation DSC. Due to the labelling differences between KiTS19 and KiTS21, all masses in KiTS19 are labelled as ‘Tumour’, whereas masses in KiTS21 are labelled as either ‘Tumour’ or ‘Cyst’.

Manual NCCT screening exhibits potential as a medium for RC early detection (O’Connor et al., 2011, 2018), yet there has been little supporting research in NCCT segmentation that may assist the automation of NCCT screening. LDCT and NCCT images are significantly noisier and less differentiated than in CECT, respectively, making target organs harder to distinguish for AI algorithms. Transference of segmentation algorithms between the CECT and NCCT or LDCT may be non-trivial due to the differences in image quality, thus new work must quantify the performance of segmentation within NCCT images, to verify the suitability of segmentation-based RC early detection in NCCT.

Classification

Renal classification algorithms generally fall into one of the following characterisations: DL-based (Han et al., 2019; Tabibu et al., 2019; Fenstermaker et al., 2020; Oberai et al., 2020; Pedersen et al., 2020; Tanaka et al., 2020; Zabihollahy et al., 2020; Uhm et al., 2021), feature analysis-based (Hodgdon et al., 2015; Schieda et al., 2015; Feng et al., 2018; Kocak et al., 2018; Lee et al., 2018; Schieda et al., 2018; Varghese et al., 2018; Erdim et al., 2020; Ma et al., 2020; Sun et al., 2020; Wang et al., 2021), or a hybrid approach (Lee et al., 2018; Tabibu et al., 2019). The higher inference time and cost of DL-based algorithms compared to feature-based algorithms is undesirable, but DL-based approaches tend to be more accurate.

DL-based classification approaches generally use ‘fine-tuned’ versions of pretrained CNN classifiers (such as ResNet, He et al., 2016; VGG, Simonyan and Zisserman, 2015; or Inception, Szegedy et al., 2016). Fine-tuning in this context means to retrain an already existing pretrained model to operate effectively in a new domain. This approach minimises the need for domain-specific labelled images (and, therefore, minimises labelling), and provides good classification performance. Feature-based algorithms operate on predetermined ROIs – image sections segmented by a radiologist or AI algorithm – and use radiomic and/or DL-derived features, that describe relationships in the local distribution of CT intensities, to classify disease.

Deep learning-based classifiers can achieve high accuracy in CT images with very little manual intervention. Tanaka et al. (2020) sought to quantify small (≤4 cm) renal mass detection accuracy in CT using axial CT slices and a fine-tuned InceptionV3 CNN; they differentiated malignant and benign masses with a maximum AUC of 0.846 in CECT and 0.562 in NCCT. Pedersen et al. (2020) trained a similar 2D slice-classifying CNN, but used it to classify each slice within each known mass’ 3D volumes to enable a slice-based voting system to differentiate patient-level RC from oncocytoma, returning a perfect validation accuracy of 100%. Han et al. (2019) sought to differentiate between clear cell RCC (ccRCC) and non-ccRCC from known RCC masses, using radiologist-selected axial CT slices from NCCT and two CECT phases, and achieved sub-type classification AUCs between 0.88 and 0.94 in an internal testing dataset.

Classification has also been performed with the following feature-based supervised learning models: support vector machines (SVM; Hodgdon et al., 2015; Schieda et al., 2015; Kocak et al., 2018; Erdim et al., 2020; Sun et al., 2020), multi-layer perceptrons (MLP; Kocak et al., 2018; Erdim et al., 2020), logistic regressions (LR; Hodgdon et al., 2015; Schieda et al., 2015; Schieda et al., 2018; Varghese et al., 2018; Ma et al., 2020; Wang et al., 2021), and decision tree methods (DT; Lee et al., 2018; Erdim et al., 2020). Some feature-based models have shown superior diagnostic performance to expert radiologists: Hodgdon et al.’s (2015) SVM-based approach classified RC in NCCT images with an AUC of around 0.85; this was much greater than the radiologists’ AUCs of 0.65 and 0.74. Sun et al.’s (2020) ‘radiologic-radiomic’ SVM model, where ‘radiologic’ refers to human-derived radiographic features and ‘radiomic’ refers to machine-derived radiographic features, differentiated RCC subtypes from benign masses. Sun et al. (2020) reported their accuracies in DSC, achieving an average of 88.3% DSC, improving upon the 78.2% average expert radiologist’s DSC (individual radiologists varied between 73.2 and 84.1%).

Across RC classification literature, the interaction between feature analysis and DL models is limited. Tabibu et al.’s (2019) classification pipeline sends patches of histopathological images to two CNNs – one CNN classifies each patch as benign/malignant, and the other generates features that are used to differentiate between RCC subtypes in a three-class SVM. In internal validation, performing classification on histopathological images, this method achieved up to 0.99 patch-wise malignancy-identification AUC, and 0.93 subtype-identification AUC. Lee et al.’s (2018) approach concatenated radiomic features with a CNN output, both evaluated over a pre-segmented ROI in a CT image and fed this concatenation to a DT classifier that differentiated angiomyolipoma without visible fat from RC with up to 0.816 AUC.

Object detection has rarely been applied to renal mass detection in CT (Yan et al., 2018; Xiong et al., 2019; Zhang et al., 2019). Zhang et al.’s (2019) renal lesion detector show a mass-level detection AUC of 0.871 in CECT; they did not compare this performance to expert radiologist performance over the same validation dataset. As in segmentation, the reduced image quality of NCCT may present issues for AI lesion detection algorithms; thus, to ensure suitability in early detection, work must be done to quantify object detection performance in NCCT.

MTL and synthetic image generation

AI has been used to support RC diagnosis in other interesting manners, including MTL and synthetic image generation (SIG). SIG aims to create new images that mimic the appearance of authentic medical images. In RC, SIG has been used to improve segmentation performance (roughly 0.5% DSC improvement, Jin et al., 2021) by synthetically expanding the size of labelled training datasets, and shows promise in improving classification performance by synthetically transferring images to more diagnostically-useful domains, such as from NCCT to CECT (Liu et al., 2020; Sassa et al., 2022). However, to the authors’ knowledge, no research has quantified the improvement in RC classification performance directly attributable to synthetic domain transfer between NCCT and CECT. MTL has been used in RC evaluation to combine learning from multiple tasks, such that they simultaneously contribute towards model training – Ruan et al. (2020) noted a 3% segmentation DSC improvement following MTL, and Pan et al. (2019) noted how classification and segmentation performance scores were both individually improved when trained together in MTL.

Alternate methods of using medical AI

Alternate detection paradigms

Rather than removing the need for pathologist personnel in screening, Gehrung et al.’s (2021) AI approach generated a proxy ‘confidence’ rating to triage patients suspected of having Barrett’s oesophagus, a precancerous state for oesophageal cancer. Their AI detected ‘indeterminate’ cases and sent these to an expert pathologist, whilst accurately assigning classifications to ‘clear’ cases. Gehrung et al.’s (2021) triage approach was rigorously assessed across multiple validation datasets and was estimated to reduce pathologist workloads by 57% without a reduction in accuracy, improving the cost-effectiveness of screening. As in Barret’s oesophagus, triaging AI may be practicable in LDCT RC screening and improve the process’ cost-effectiveness (Wilson–Junger criterion 8, Table 1).

Khosravan et al. (2019) found that humans tend to have higher specificity and AI algorithms tend to have higher sensitivity in NCCT lung cancer detection; in response, they constructed a ‘complimentary’ computer-aided diagnosis system to bridge the performance gap between radiologists and AI. Khosravan et al.’s (2019) system let a radiologist evaluate an input NCCT image as the AI system segmented and classify each gaze-deduced region of interest, generated by the radiologist’s eye movement, automatically. This study failed to specify the improvement in cancer detection, or workload reduction, directly attributable to their software, instead plainly evaluated the performance of segmentation (91% DSC) and classification (97% accuracy – AUC not reported).

Object detection in AI cancer detection

Ardila et al. (2019) used an object-detection algorithm to identify lung nodules in NCCT with high accuracy, allowing patient-level early cancer detection AUC of 0.944. Welikala et al. (2020) used an object detection algorithm to identify oral lesions in plain photographic images of the oral cavity, allowing patient-level cancer classification, and achieving a patient-level classification DSC between 78 and 87% (AUC not reported). Nguyen et al. (2022) proposed a circular ‘bounding-box’ object detection algorithm for general biological purposes, as certain biological structures tend to be more circular/spherical than rectangular/cuboidal such as cells, masses, and some organs. They proved that their ‘CircleNet’ object-detection algorithm showed overall superior performance to other state-of-the-art algorithms in detecting nuclei and glomeruli.

Synthetic image generation

Santini et al.’s (2018) DL workflow synthetically enhance NCCT images, promoting them to pseudo-CECT, to enable accurate estimation of patient cardiac volumes. Santini et al. (2018) proved the efficacy of this method by highlighting the segmentation improvement associated with synthetic CECT generation; their framework, performing segmentation over synthetic CECTs, was more accurate than a human over an equivalent set of NCCTs (DSC of 0.89 and 0.85, respectively). Hu et al. (2022) built a generative adversarial network (GAN) to generate realistic synthetic CECT images that improve the conspicuity of abdominal aortic aneurysms in NCCT images. Their GAN made use of U-Net to generate synthetic CECT images, and was trained in MTL – using vascular structure segmentation as an auxiliary task to boost the performance of CECT generation. Hu et al. (2022) found that their GAN outperformed stand-alone U-Net, and other SIG algorithms such as pix2pix (Isola et al., 2017) and MW-CNN (Liu et al., 2018), in terms of average validation error and signal-to-noise ratio. Qualitatively, Hu et al. (2022) showed clearly that the noise produced in U-Net-based NCCT to CECT translation is minimised by its incorporation into a GAN. Hu et al. (2022) did not directly quantify the improvement in aneurysm detection directly attributable to their synthetic CT enhancement, but they did determine case-level aneurysm detection DSC to be 85%.

Emergent ideas across AI and computer vision

Segmentation

Yang et al. (2022) found that exhaustive hyperparameter optimisation of large AI models, such as CNNs and transformers, is possible – they showed neural networks over a very large range of sizes can share common optimal hyperparameters if they are initialised ‘correctly’. This correct initialisation allows grid-search-based objective hyperparameter optimisation, which nnU-Net established as primarily important in segmentation. Also, the intrinsic locality of convolutional operations in CNNs may limit U-Net’s performance in segmentation tasks with global pattern dependencies. Introducing transformers, capable of global attention and understanding the relationships between all input data, to the U-Net architecture may allow the model to ‘see’ much larger volumes during segmentation, which may improve segmentation accuracy. TransU-Net and UNETR both implemented transformers into U-Net’s CNN architecture and significantly improved upon U-Net’s segmentation performance in multi-organ segmentation tasks (Chen et al., 2021; Hatamizadeh et al., 2022).

Classification

Following the introduction of transformers (Vaswani et al., 2017; Dosovitskiy et al., 2020), a new generation of state-of-the-art classifiers (including ConvNeXt, Liu et al., 2022), Swin (Liu et al., 2021) and CoaT (Xu et al., 2021), have superseded the commonly used CNNs Resnet, VGG and Inception in terms of ImageNet classification accuracy. This new generation shows improved performance over the same tasks due to their new training regimes, new hyperparameters and new architectures. ConvNeXT (which, like the previous generation of classifiers, is a pure CNN) tweaked its properties to take advantage of insights made by transformers models (Liu et al., 2022) and shows improved performance over the previous generation without incurring greater cost during inference.

Multi-task learning

Standley et al. (2020) assessed various methods of combining AI training regimes. They found that some ‘complex’ tasks, such as segmentation, require greater number of training samples for optimal performance than other ‘simpler’ tasks, and that these more complex tasks’ performances would suffer if paired with a simple task in MTL. Standley et al. (2020) also found that some tasks seemed to consistently act as ‘auxiliaries’ – boosting the learning performance of the network for other tasks without ever performing significantly well themselves in MTL. Despite these findings, they found that the relationships between task pairings – that is, the tendency of tasks to help or hinder each other’s training during MTL – was not independent of the training setup, meaning MTL relationships between tasks cannot be completely generalised across models with distinct network architectures, hyperparameters, and training data.

Discussion

Renal segmentation has the potential in assisting RC diagnosis – for example, accurately delineating tumour regions enables feature-based classification, which shows comparable, or superior, diagnostic performance to expert radiologists. Maximising renal segmentation accuracy in LDCT may enable accurate feature-based classification methods to be applied in LDCT early detection automatically, removing much of the manual labour of RC screening. High accuracy is essential in early detection methods; thus, given the accuracy of the feature-based classification methods in NCCT imaging (as in Hodgdon et al., 2015), a high-accuracy renal segmentation method for LDCT is likely to enable RC early detection screening.

Whilst nnU-Net established the primacy of hyperparameter optimisation in segmentation performance, it does not provide a framework for hyperparameter optimisation itself, instead relying on experimentally derived heuristics for hyperparameter selection. Using Yang et al.’s (2022) ‘maximal parameter update’ hyperparameter optimisation allows a definitive optimisation of any CNN or transformer, which should improve upon nnU-Net’s heuristics-led approach. Also, despite nnU-Net’s state-of-the-art inter-domain performance, the intrinsic locality of convolutional operations in U-Net’s purely convolutional architecture may limit its segmentation performance. Introducing transformers to U-Net’s architecture, as in TransU-Net, enables global attention mechanisms that may improve RC segmentation accuracy over a whole NCCT volume. Applying transformer-informed segmentation methods like TransU-Net, and objectively optimising its hyperparameters using ‘maximal parameter updates’ may improve RC segmentation performance over existing nnU-Net-led approaches.

Given the potential for RC early detection in LDCT, there is a need for more research quantifying RC segmentation performance in LDCT. Investigations into general NCCT segmentation have shown that using synthetic contrast enhancement as an auxiliary training task in MTL can improve segmentation accuracy. Therefore, an investigation in renal LDCT segmentation may be improved by introducing synthetic enhancement to CECT as an auxiliary learning task in MTL. Such an investigation would likely be complicated by Standley et al. (2020) findings – that MTL task relationships can be unique to each configuration of network architecture, hyperparameters, and dataset domain.

Like segmentation, the lack of research quantifying RC object detection performance in LDCT represents a gap in the literature. Object detection and classification performance could be improved by the introduction of the new generation transformer-inspired classifiers that consistently show higher classification accuracies than their predecessors. Also, assessing the MTL relationship between classification, segmentation, and object detection in RC early detection may lead to improved mass detection, and therefore early detection, performance.

Pedersen et al.’s (2020) and Gehrung et al.’s (2021) approach of generating an image-based intra-patient biomarker voting system may be applicable to RC early detection. Both Pedersen et al. (2020) and Gehrung et al. (2021) evaluated biomarker presence in fractionated tiles of input images and used the ratio of biomarker-positive to biomarker-negative tiles to classify the inputs, leading to high-accuracy results in validation. Applying an analogous approach, using the new generation of classifiers, to the early detection of RC masses in LDCT could enable highly robust automated triaging, or diagnosis, for RC early detection screening programmes.

Conclusion

This manuscript highlights and summarises existing AI method in RC diagnosis and suggests how these can be repurposed to enable RC early detection. After summarising existing segmentation, classification, and other AI methods in RC diagnosis, a review of analogous cancer detection and diagnosis methods across broader cancer literature and computer vision was conducted. Contrasting the RC-specific workflows to their equivalents across computer vision and other cancer domains allowed the generation of novel RC-specific research proposals that may enable AI-based RC early detection.

Open peer review

To view the open peer review materials for this article, please visit http://doi.org/10.1017/pcm.2022.9.

Supplementary material

For supplementary material accompanying this paper visit https://doi.org/10.1017/pcm.2022.9.

S2752614322000096sup001.docx (29.4KB, docx)

click here to view supplementary material

Financial support

This work was supported by the International Alliance for Cancer Early Detection, a partnership between Cancer Research UK (C14478/A27855), Canary Center at Stanford University, the University of Cambridge, OHSU Knight Cancer Institute, University College London and the University of Manchester. This work was also supported by the CRUK National Cancer Imaging Translational Accelerator (NCITA) (C42780/A27066), and The Mark Foundation for Cancer Research and Cancer Research UK (CRUK) Cambridge Centre (C9685/A25177). Additional support has been provided by the Wellcome Trust Innovator Award, UK (215733/Z/19/Z) and the National Institute of Health Research (NIHR) Cambridge Biomedical Research Centre (BRC-1215-20014). The views expressed are those of the authors and not necessarily those of the NHS, the NIHR or the Department of Health and Social Care.

Competing interest

The authors of this manuscript declare relationships with the following companies: E.S. is a co-founder and shareholder of Lucida Medical Ltd. L.E.S. has received consulting fees from Lucida Medical Ltd. The remaining authors declare that they have no conflicts of interest to declare.

References

  1. Ardila D, Kiraly AP, Bharadwaj S, Choi B, Reicher JJ, Peng L, Tse D, Etemadi M, Ye W, Corrado G, Naidich DP and Shetty S (2019) End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nature Medicine 25(6), 954–961. 10.1038/s41591-019-0447-x. [DOI] [PubMed] [Google Scholar]
  2. Beinfeld MT, Wittenberg E and Gazelle GS (2005) Cost-effectiveness of whole-body CT Screening1. Radiology 234(2), 415. [DOI] [PubMed] [Google Scholar]
  3. Black WC, Gareen IF, Soneji SS, Sicks JD, Keeler EB, Aberle DR, Naeim A, Church TR, Silvestri GA, Gorelick J and Gatsonis C (2014) Cost-effectiveness of CT screening in the National Lung Screening Trial. New England Journal of Medicine 371(19), 1793–1802. 10.1056/nejmoa1312547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Challenge Leaderboard (2019) KiTS19 Grand Challenge. Available at https://kits19.grand-challenge.org/evaluation/challenge/leaderboard/?date=2019-08-07 (accessed 7 March 2019).
  5. Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, Lu L, Yuille AL and Zhou Y (2021) TransUNet: Transformers make strong encoders for medical image segmentation. Preprint, arXiv:2102.04306.
  6. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J and Houlsby N (2020) An image is worth 16×16 words: Transformers for image recognition at scale. Preprint, arXiv:2010.11929.
  7. Erdim C, Yardimci AH, Bektas CT, Kocak B, Koca SB, Demir H and Kilickesmez O (2020) Prediction of benign and malignant solid renal masses: Machine learning-based CT texture analysis. Academic Radiology 27(10), 1422–1429. [DOI] [PubMed] [Google Scholar]
  8. Feng Z, Rong P, Cao P, Zhou Q, Zhu W, Yan Z, Liu Q and Wang W (2018) Machine learning-based quantitative texture analysis of CT images of small renal masses: Differentiation of angiomyolipoma without visible fat from renal cell carcinoma. European Radiology 28(4), 1625–1633. 10.1007/s00330-017-5118-z. [DOI] [PubMed] [Google Scholar]
  9. Fenstermaker M, Tomlins SA, Singh K, Wiens J and Morgan T (2020) Development and validation of a deep-learning model to assist with renal cell carcinoma histopathologic interpretation. Urology 144, 152–157. [DOI] [PubMed] [Google Scholar]
  10. Fitzmaurice C, Abate D, Abbasi N, Abbastabar H, Abd-Allah F, Abdel-Rahman O, Abdelalim A, Abdoli A, Abdollahpour I, Abdulle ASM, Abebe ND, Abraha HN, Abu-Raddad LJ, Abualhasan A, Adedeji IA, Advani SM, Afarideh M, Afshari M, Aghaali M, and Aghaali M. (2019) Global, regional, and national cancer incidence, mortality, years of life lost, years lived with disability, and disability-adjusted life-years for 29 cancer groups, 1990 to 2017. JAMA Oncology 5(12), 1749. 10.1001/jamaoncol.2019.2996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Freer-Smith C, Harvey-Kelly L, Mills K, Harrison H, Rossi SH, Griffin SJ, Stewart GD and Usher-Smith JA (2021) Reasons for intending to accept or decline kidney cancer screening: Thematic analysis of free text from an online survey. BMJ Open 11(5), e044961. 10.1136/bmjopen-2020-044961. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Gehrung M, Crispin-Ortuzar M, Berman AG, O’Donovan M, Fitzgerald RC and Markowetz F (2021) Triage-driven diagnosis of Barrett’s esophagus for early detection of esophageal adenocarcinoma using deep learning. Nature Medicine 27(5), 833–841. 10.1038/s41591-021-01287-9. [DOI] [PubMed] [Google Scholar]
  13. Guidelines for the Management of Renal Cancer (2016) West Midlands Expert Advisory Group for Urological Cancer. Available at https://www.england.nhs.uk/mids-east/wp-content/uploads/sites/7/2018/05/guidelines-for-the-management-of-renal-cancer.pdf (accessed 15 May 2022).
  14. Han S, Hwang SI and Lee HJ (2019) The classification of renal cancer in 3-phase CT images using a deep learning method. Journal of Digital Imaging 32(4), 638–643. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Harvey-Kelly LLW, Harrison H, Rossi SH, Griffin SJ, Stewart GD and Usher-Smith JA (2020) Public attitudes towards screening for kidney cancer: An online survey. BMC Urology 20(1), 170. 10.1186/s12894-020-00724-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, Roth HR and Xu D (2022) UNETR: Transformers for 3D medical image segmentation. In 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Los Alamitos, CA, USA. IEEE. pp. 1748–1758.doi: 10.1109/WACV51458.2022.00181K. [DOI] [Google Scholar]
  17. He K, Zhang X, Ren S and Sun J (2016) Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, IEEE, pp. 770–778 doi: 10.1109/CVPR.2016.90. [DOI] [Google Scholar]
  18. Heller N, Sathianathen N, Kalapara A, Walczak E, Moore K, Kaluzniak H, Rosenberg J, Blake P, Rengel Z, Oestreich M, Dean J, Tradewell M, Shah A, Tejpaul R, Edgerton Z, Peterson M, Raza S, Regmi S, Papanikolopoulos N and Weight C (2019) The KiTS19 challenge data: 300 kidney tumor cases with clinical context, CT semantic segmentations, and surgical outcomes. Preprint, arXiv:1904.00445.
  19. Hodgdon T, Matthew DFMI, Schieda N, Flood TA, Lamb L and Thornhill RE (2015) Can quantitative CT texture analysis be used to differentiate fat-poor renal angiomyolipoma from renal cell carcinoma on unenhanced CT images? Radiology 276(3), 787–796. [DOI] [PubMed] [Google Scholar]
  20. Hu T, Oda M, Hayashi Y, Lu Z, Kumamaru KK, Akashi T, Aoki S and Mori K (2022) Aorta-aware GAN for non-contrast to artery contrasted CT translation and its application to abdominal aortic aneurysm detection. International Journal of Computer Assisted Radiology and Surgery 17(1), 97–105. 10.1007/s11548-021-02492-0. [DOI] [PubMed] [Google Scholar]
  21. Hunink MGM and Gazelle GS (2003) CT screening: A trade-off of risks, benefits, and costs. Journal of Clinical Investigation 111(11), 1612–1619. 10.1172/jci18842. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Isensee F, Jaeger PF, Simon AAK, Petersen J and Maier-Hein KH (2021) nnU-net: A self-configuring method for deep learning-based biomedical image segmentation. Nature Methods 18, 203–211. [DOI] [PubMed] [Google Scholar]
  23. Isensee F and Maier-Hein KH (2019) An attempt at beating the 3D U-Net. Preprint, arXiv:1908.02182.
  24. Ishikawa S, Aoki J, Ohwada S, Takahashi T, Morishita Y and Ueda K (2007) Mass screening of multiple abdominal solid organs using Mobile helical computed tomography scanner—A preliminary report. Asian Journal of Surgery 30(2), 118–121. 10.1016/s1015-9584(09)60143-3. [DOI] [PubMed] [Google Scholar]
  25. Isola P, Zhu J-Y, Zhou T and Efros AA (2017) Image-to-image translation with conditional adversarial networks. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI, USA, IEEE. [Google Scholar]
  26. Jensen MD, Siersma V, Rasmussen JF and Brodersen J (2020) Direct and indirect healthcare costs of lung cancer CT screening in Denmark: A registry study. BMJ Open 10(1), e031768. 10.1136/bmjopen-2019-031768. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Jin Q, Cui H, Sun C, Meng Z and Su R (2021) Free-form tumor synthesis in computed tomography images via richer generative adversarial network. Knowledge-Based Systems 218, 106753. [Google Scholar]
  28. Khosravan N, Celik H, Turkbey B, Jones EC, Wood B and Bagci U (2019) A collaborative computer aided diagnosis (C-CAD) system with eye-tracking, sparse attentional model, and deep learning. Medical Image Analysis 51, 101–115. 10.1016/j.media.2018.10.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. KiTS21 (2021) KiTS21 results. Available at https://kits21.kits-challenge.org/results (accessed 7 March 2021).
  30. Kocak B, Yardimci AH, Bektas CT, Turkcanoglu MH, Erdim C, Yucetas U, Koca SB and Kilickesmez O (2018) Textural differences between renal cell carcinoma subtypes: Machine learning-based quantitative computed tomography texture analysis with independent external validation. European Journal of Radiology 107, 149–157. [DOI] [PubMed] [Google Scholar]
  31. Lee H, Hong H, Kim J and Jung DC (2018) Deep feature classification of angiomyolipoma without visible fat and renal cell carcinoma in abdominal contrast-enhanced CT images with texture image patches and hand-crafted feature concatenation. Medical Physics 45(4), 1550–1561. [DOI] [PubMed] [Google Scholar]
  32. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S and Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 2021 pp. 9992–10002. doi: 10.1109/ICCV48922.2021.00986 [DOI] [Google Scholar]
  33. Liu Z, Mao H, Wu C-Y, Feichtenhofer C, Darrell T and Xie S (2022) A ConvNet for the 2020s. Preprint, arXiv:2201.03545v2.
  34. Liu J, Tian Y, Ağıldere AM, Haberal KM, Coşkun M, Duzgol C and Akin O (2020) DyeFreeNet: Deep virtual contrast CT synthesis. International Workshop on Simulation and Synthesis in Medical Imaging, SASHIMI 2020: Simulation and Synthesis in Medical Imaging pp 80–89, Springer International Publishing. 10.1007/978-3-030-59520-3_9. [DOI] [Google Scholar]
  35. Liu P, Zhang H, Zhang K, Lin L and Zuo W (2018) Multi-level wavelet-CNN for image restoration. In IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). Salt Lake City, UT, USA, IEEE. [Google Scholar]
  36. Ljungberg B, Bensalah K, Canfield S, Dabestani S, Hofmann F, Hora M, Kuczyk MA, Lam T, Marconi L, Merseburger AS, Mulders P, Powles T, Staehler M, Volpe A and Bex A (2015) EAU guidelines on renal cell carcinoma: 2014 update. European Urology 67(5), 913–924. 10.1016/j.eururo.2015.01.005. [DOI] [PubMed] [Google Scholar]
  37. Ma Y, Cao F, Xu X and Ma W (2020) Can whole-tumor radiomics-based CT analysis better diferentiate fat-poor angiomyolipoma from clear cell renal cell caricinoma: Compared with conventional CT analysis? Abdominal Radiology 45, 2500–2507. [DOI] [PubMed] [Google Scholar]
  38. Nguyen EH, Yang H, Deng R, Lu Y, Zhu Z, Roland JT, Lu L, Landman BA, Fogo AB and Huo Y (2022) Circle representation for medical object detection. IEEE Transactions on Medical Imaging 41(3), 746–754. 10.1109/tmi.2021.3122835. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. NLST (2011) Reduced lung-cancer mortality with low-dose computed tomographic screening. New England Journal of Medicine 365(5), 395–409. 10.1056/nejmoa1102873. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Oberai A, Varghese B, Cen S, Angelini T, Hwang D, Gill I, Aron M, Lau C and Duddalwar V (2020) Deep learning based classification of solid lipid-poor contrast enhancing renal masses using contrast enhanced CT. The British Journal of Radiology 93(1111), 20200002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. O’Connor SD, Pickhard PJ, Kim DH, Oliva MR and Silverman SG (2011) Incidental finding of renal masses at unenhanced CT: Prevalence and analysis of features for guiding management. American Journal of Roentgenology 197(1), 139–145. [DOI] [PubMed] [Google Scholar]
  42. O’Connor SD, Silverman SG, Cochon LR and Khorasani RK (2018) Renal cancer at unenhanced CT: Imaging features, detection rates, and outcomes. Abdominal Radiology 43(7), 1756–1763. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Pan T, Shu H, Coatrieux J-L, Yang G, Wang C, Lu Z, Zhou Z, Kong Y, Tang L, Zhu X and Dillenseger J-L (2019) A multi-task convolutional neural network for renal tumor segmentation and classification using multi-phasic CT images, IEEE International Conference on Image Processing (ICIP), 2019, pp. 809–813, doi: 10.1109/ICIP.2019.8802924. [DOI]
  44. Pedersen M, Andersen MB, Christiansen H and Azawi NH (2020) Classification of renal tumour using convolutional neural networks to detect oncocytoma. European Journal of Radiology 133, 109343. [DOI] [PubMed] [Google Scholar]
  45. Rabjerg M, Mikkelsen MN, Walter S and Marcussen N (2014) Incidental renal neoplasms: Is there a need for routine screening? A Danish single-center epidemiological study. APMIS 122(8), 708–714. 10.1111/apm.12282. [DOI] [PubMed] [Google Scholar]
  46. Ronneberger O, Fischer P and Brox T (2015) U-Net: Convolutional Networks for Biomedical Image Segmentation. In Lecture Notes in Computer Science. 9351, 234–241. Springer, 10.1007/978-3-319-24574-4_28 [DOI] [Google Scholar]
  47. Rossi SH, Klatte T, Usher-Smith J and Stewart GD (2018) Epidemiology and screening for renal cancer. World Journal of Urology 36(9), 1341–1353. 10.1007/s00345-018-2286-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Ruan Y, Li D, Marshall H, Miao T, Cossetto T, Chan I, Daher O, Accorsi F, Goela A and Li S (2020) MB-FSGAN: Joint segmentation and quantification of kidney tumor on CT by the multi-branch feature sharing generative adversarial network. Medical Image Analysis 64, 101721. [DOI] [PubMed] [Google Scholar]
  49. Santini G, Zumbo LM, Martini N, Valvano G, Leo A, Ripoli A, Avogliero F, Chiappino D and Latta DD (2018) Synthetic contrast enhancement in cardiac CT with deep learning. Preprint, arXiv:1807.01779.
  50. Sassa N, Kameya Y, Takahashi T, Matsukawa Y, Majima T, Tsuruta K, Kobayashi I, Kajikawa K, Kawanishi H, Kurosu H, Yamagiwa S, Takahashi M, Hotta K, Yamada K and Yamamoto T (2022) Creation of synthetic contrast-enhanced computed tomography images using deep neural networks to screen for renal cell carcinoma. Cold Spring Harbor Laboratory. 10.1101/2022.01.12.22269120. [DOI] [PMC free article] [PubMed]
  51. Schieda N, Lim RS, Krishna S, Matthew DFMI, Flood TA and Thornhill RE (2018) Diagnostic accuracy of unenhanced CT analysis to differentiate low-grade from high-grade Chromophobe renal cell carcinoma. American Journal of Roentgenology 210, 1079–1087. [DOI] [PubMed] [Google Scholar]
  52. Schieda N, Thornhill RE, Al-Subhi M, Matthew DFMI, Shabana WM, van der Pol CB and Flood TA (2015) Diagnosis of Sarcomatoid renal cell carcinoma with CT: Evaluation by qualitative imaging features and texture analysis. American Journal of Roentgenology 204(5), 1013–1023. [DOI] [PubMed] [Google Scholar]
  53. Simonyan K and Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. 3rd International Conference on Learning Representations (ICLR 2015), 1–14.
  54. Standley T, Zamir AR, Chen D, Guibas L, Malik J and Savarese S (2020) Which tasks should be Learned Together in Multi-task Learning? In International Conference on Machine Learning, https://arxiv.org/abs/1905.07553. [Google Scholar]
  55. Stewart GD (2021) Yorkshire kidney screening trial, ISRCTN18055040. Available at 10.1186/ISRCTN18055040, accessed on 29/05/2022. [DOI]
  56. Sun X-Y, Feng Q-X, Xu X, Zhang J, Zhu F-P, Yang Y-H and Zhang Y-D (2020) Radiologic-radiomic machine learning models for differentiation of benign and malignant solid renal masses. Comparison With Expert-Level Radiologists 214(1), 44–54. [DOI] [PubMed] [Google Scholar]
  57. Szegedy C, Vanhoucke V, Ioffe S, Shlens J and Wojna Z (2016) Rethinking the inception architecture for computer vision. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA. IEEE. [Google Scholar]
  58. Tabibu S, Vinod PK and Jawahar CV (2019) Pan-renal cell carcinoma classification and survival prediction from histopathology images using deep learning. Scientific Reports 9, 10509. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Tanaka T, Huang Y, Marukawa Y, Tsuboi Y, Masaoka Y, Kojima K, Iguchi T, Hiraki T, Gobara H, Yanai H, Nasu Y and Kanazawa S (2020) Differentiation of small (≤4 cm) renal masses on multiphase contrast-enhanced CT by deep learning. American Journal of Roentgenology 214, 605–612. [DOI] [PubMed] [Google Scholar]
  60. Uhm K-H, Jung S-W, Choi MH, Shin H-K, Yoo J-I, Oh SW, Kim JY, Kim HG, Lee YJ, Youn SY, Hong S-H and Ko S-J (2021) Deep learning for end-to-end kidney cancer diagnosis on multi-phase abdominal computed tomography. NPJ Precision Oncology 5, 54. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Varghese BA, Chen F, Hwang DH, Cen SY, Desai B, Gill IS and Duddalwar VA (2018) Differentiation of predominantly solid enhancing lipid-poor renal cell masses by use of contrast-enhanced CT: Evaluating the role of texture in tumor subtyping. American Journal of Roentgenology 211, W288–W296. [DOI] [PubMed] [Google Scholar]
  62. Vasudev NS, Wilson M, Stewart GD, Adeyoju A, Cartledge J, Kimuli M, Datta S, Hanbury D, Hrouda D, Oades G, Patel P, Soomro N, Sullivan M, Webster J, Selby PJ and Banks RE (2020) Challenges of early renal cancer detection: Symptom patterns and incidental diagnosis rate in a multicentre prospective UK cohort of patients presenting with suspected renal cancer. BMJ Open 10(5), e035938. 10.1136/bmjopen-2019-035938. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L and Polosukhin I (2017) Attention is all you need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017 December 4–9, 2017, Long Beach, CA, USA. Curran Associates Inc. https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf [Google Scholar]
  64. Volpe A, Panzarella T, Rendon RA, Haider MA, Kondylis FI and Jewett MAS (2004) The natural history of incidentally detected small renal masses. Cancer 100(4), 738–745. 10.1002/cncr.20025. [DOI] [PubMed] [Google Scholar]
  65. Wang X, Song G and Jiang H (2021) Differentiation of renal angiomyolipoma without visible fat from small clear cell renal cell carcinoma by using specific region of interest on contrast-enhanced CT: A new combination of quantitative tools. Cancer Imaging 21, 47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Welikala RA, Remagnino P, Lim JH, Chan CS, Rajendran S, Kallarakkal TG, Zain RB, Jayasinghe RD, Rimal J, Kerr AR, Amtha R, Patil K, Tilakaratne WM, Gibson J, Cheong SC and Barman SA (2020) Automated detection and classification of oral lesions using deep learning for early detection of oral cancer. IEEE Access 8, 132677–132693. 10.1109/access.2020.3010180. [DOI] [Google Scholar]
  67. Wilson JMG, Jungner G and World Health Organization (1968) Principles and Practice of Screening for Disease. Wilson J. M. G., Jungner G.. World Health Organization. https://apps.who.int/iris/handle/10665/37650. [Google Scholar]
  68. Xiong Z, Zhang H, Chen Y and Song Y (2019) Deep Ensemble Learning Network for Kidney Lesion Detection. Chinese Automation Congress (CAC). pp. 3841–3846, doi: 10.1109/CAC48633.2019.8997272. [DOI] [Google Scholar]
  69. Xu W, Xu Y, Chang T and Tu Z (2021) Co-scale Conv-attentional image transformers. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, IEEE. [Google Scholar]
  70. Yan K, Wang X, Lu L and Summers RM (2018) DeepLesion: Automated mining of large-scale lesion annotations and universal lesion detection with deep learning. Journal of Medical Imaging 5(03), 1. 10.1117/1.jmi.5.3.036501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Yang G, Hu EJ, Babuschkin I, Sidor S, Liu X, Farhi D, Ryder N, Pachocki J, Chen W and Gao J (2022) Tensor programs V: Tuning large neural networks via zero-shot hyperparameter transfer. In Advances in Neural Information Processing Systems, p. 34. arXiv, 10.48550/ARXIV.2203.03466. https://arxiv.org/abs/2203.03466 [DOI] [Google Scholar]
  72. Zabihollahy F, Schieda N, Krishna S and Ukwatta E (2020) Automated classification of solid renal masses on contrast-enhanced computed tomography images using convolutional neural network with decision fusion. European Radiology 30, 5183–5190. [DOI] [PubMed] [Google Scholar]
  73. Zhang H, Chen Y, Song Y, Xiong Z, Yang Y and Jonathan Wu QM (2019) Automatic kidney lesion detection for CT images using morphological cascade convolutional neural networks. IEEE Access 7, 83001–83011. 10.1109/access.2019.2924207. [DOI] [Google Scholar]
  74. Zhang L, Yao L, Li X, Jewett MAS, He Z and Zhou L (2016) Natural history of renal cell carcinoma: An immunohistochemical analysis of growth rate in patients with delayed treatment. Journal of the Formosan Medical Association 115(6), 463–469. [DOI] [PubMed] [Google Scholar]
  75. Zhao Z, Chen H and Wang L (2022) A coarse-to-fine framework for the 2021 kidney and kidney tumor segmentation challenge. International Challenge on Kidney and Kidney Tumor Segmentation, KiTS 2021: Kidney and Kidney Tumor Segmentation pp 53–58. Springer International Publishing, 10.1007/978-3-030-98385-7_8. [DOI] [Google Scholar]
Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr2

Review: The environmental impact of data-driven precision medicine initiatives — R0/PR2

Han Van Krieken 1

Comments to Author: This review deals with advances in AI for radiological early detection of renal cell carcinoma (RCC). The focus is on technical aspects and the developments therein, whereas the route to implementation and the role it can have in screening is superficially dealt with. The title is not fully in line with the scope.

This review is interesting mainly for readers who are interested in technical aspects of the use of AI in radiology

Remarks:

- although this is not a systematic review, still some indication on the approach to find and select articles are needed

- the authors describe that many of the 10 Wilson-Junger criteria are met, yet give examples that are not met. I am doubtful that CT for early RCC is really close to implementation; a table might be helpful

- it is stated that for screening te procedure needs to quick, with rapid reporting: for screening this is less important for clinical questions; in fact some actual screening methods like for colorectal cancer and cervical cancer are not that quick

- the term classification is used in two different situations: radiological and pathological. This leads to confusion: in the one case it is tissue/tumor separation in the other it is the categorization into tumortype

- some examples of application of AI in clinical practice are given, including in pathology. Although there are several articles, there is very little implementation in pathology practice, in fact probably only in the field of lymph node evaluation for metastasis, which is not mentioned in the review. Furthermore, there is literature on the use of AI in radiology in lung and breast cancer screening that gets very little attention

Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr3

Recommendation: The environmental impact of data-driven precision medicine initiatives — R0/PR3

Editor: Damian Smedley1

Comments to Author: This manuscript gives an extensive and technical overview of repurposing existing AI approaches for RCC early detection, ending with recommendations to improve both segmentation and classification approaches to enable early RCC detection.

This is a well-written review of the literature and I believe will be well received by the community.

A few suggestions:

1) I suggest combining subsections 3.3 and 3.4 to be consistent with section 2.

2) I suggest including a table summarising each reference referred to in sections 3 and 4 would support the reader in navigating between referrals to the references in the Discussion with the main body of the text

3) I wonder if the title should be something like: 'ADVANCING EARLY DETECTION OF RENAL CANCER WITH ARTIFICIAL INTELLIGENCE IN COMPUTED TOMOGRAPHY'. As not all approaches in the paper are 'new'?

4) The abstract of the article and impact statement must be given in the article before the introduction.

In general, the paper was clear and adds to the knowledge in this field - I hope to see the recommendations made in this article taken forward.

Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr4

Decision: The environmental impact of data-driven precision medicine initiatives — R0/PR4

Editor: Anna Dominiczak1

No accompanying comment.

Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr6

Review: The environmental impact of data-driven precision medicine initiatives — R1/PR6

Han Van Krieken 1

Comments to Author: This review has improved substantially

Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr7

Recommendation: The environmental impact of data-driven precision medicine initiatives — R1/PR7

Editor: Damian Smedley1

No accompanying comment.

Camb Prism Precis Med. doi: 10.1017/pcm.2022.1.pr8

Decision: The environmental impact of data-driven precision medicine initiatives — R1/PR8

Editor: Anna Dominiczak1

No accompanying comment.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    For supplementary material accompanying this paper visit https://doi.org/10.1017/pcm.2022.9.

    S2752614322000096sup001.docx (29.4KB, docx)

    click here to view supplementary material


    Articles from Cambridge Prisms: Precision Medicine are provided here courtesy of Cambridge University Press

    RESOURCES