Skip to main content
BioMedical Engineering OnLine logoLink to BioMedical Engineering OnLine
. 2024 Mar 5;23:27. doi: 10.1186/s12938-024-01209-z

Application of visual transformer in renal image analysis

Yuwei Yin 1,2, Zhixian Tang 2,, Huachun Weng 1,2,
PMCID: PMC10913284  PMID: 38439100

Abstract

Deep Self-Attention Network (Transformer) is an encoder–decoder architectural model that excels in establishing long-distance dependencies and is first applied in natural language processing. Due to its complementary nature with the inductive bias of convolutional neural network (CNN), Transformer has been gradually applied to medical image processing, including kidney image processing. It has become a hot research topic in recent years. To further explore new ideas and directions in the field of renal image processing, this paper outlines the characteristics of the Transformer network model and summarizes the application of the Transformer-based model in renal image segmentation, classification, detection, electronic medical records, and decision-making systems, and compared with CNN-based renal image processing algorithm, analyzing the advantages and disadvantages of this technique in renal image processing. In addition, this paper gives an outlook on the development trend of Transformer in renal image processing, which provides a valuable reference for a lot of renal image analysis.

Keywords: Deep learning, Transformer, Convolutional neural network, Attention mechanism, Kidney disease

Background

Kidney disease is a series of infections caused by kidney damage in function, morphology, or structure. Common kidney diseases include glomerulonephritis, pyelonephritis, diabetic nephropathy, hypertensive nephropathy, kidney stones, etc. Glomerulonephritis and diabetic nephropathy are the leading causes of chronic kidney failure. Today, ten percent of the world's population suffers from chronic kidney disease (CKD), which has become one of the most prevalent and fatal diseases and seriously affects people's health [1]. Kidney stones disease (KSD) is a common disease caused by solid mineral deposits that form in the kidneys [2]. According to the World Health Organization, approximately 5–10% of the global adult population suffers from kidney stones, with 10% and 14% in some developed countries in Europe and North America, respectively [3]. Meanwhile, kidney stones have been on the rise in the past decades. Renal cancer is a common urological malignancy, with more than 4 million new cases diagnosed yearly [4]. Therefore, improving the accuracy of diagnosis and early detection rate of nephrolithiasis is very important for the treatment and prognosis of patients.

With the development of digital medical technology, medical image processing technology has also been rapidly developed and has become one of the crucial tools in the medical field, especially in diagnosing renal diseases. Several medical imaging modalities exist, such as ultrasonography, computed tomography (CT) [5], and magnetic resonance imaging (MRI) [6]. However, imaging tests may require longer scanning times, and diagnostic images need more time and effort from healthcare professionals. Long-term fatigue of healthcare workers is likely to result in subjective misdiagnosis or underdiagnosis.

Some studies have shown that using machine learning in medical imaging can reduce the possibility of diagnostic errors and thus effectively improve diagnostic accuracy [2]. Therefore, improving the ability and automation of image analysis is a widespread issue in medical research today. Deep learning, as a branch of machine learning, has been tried to be applied in diagnosing CKD and predicting the decline of renal function [7], renal insufficiency, and diabetic nephropathy.

Deep Self-Attention Network (Transformer), as a new type of sequence model, has been widely recognized for its excellent performance in fields such as natural language processing [8]. Kidney CT/MRI images are sequential structural data with complex structural correlations between different parts. The transformer can simultaneously learn the contextual information of other parts of kidney images through the mechanism of multi-attention and capture the global structural relationship of the images more comprehensively and accurately to improve the recognition effect. The focus of current research is how to introduce it into medical image processing, especially in kidney disease. Moreover, the Transformer framework is more general, and the trained base model can be used for other renal image analysis tasks, such as classification, detection, segmentation, etc. This paper outlines the current stage of the Transformer's application in kidney image classification, segmentation, and detection and compares it with traditional CNN models.

Introduction to transformer

The Transformer model is the first transduction model that relies exclusively on self-attention to compute its input and output representations without recurrent neural networks (RNNs) or CNNs for sequence comparison [8]. Compared to commonly used models such as RNNs and CNNs, Transformer has a higher parallel computation capability due to an attentional mechanism that simultaneously allows the computation to consider all input words or characters. Moreover, the self-attention mechanism can effectively handle long sequential data and improve the modeling ability of long-range dependencies. The transformer abstracts the encoder and decoder into individual modules (as shown in Fig. 1). In the encoder, the inputs are mapped to a multidimensional space, and the input representation is learned through the multi-head self-attention mechanism. The feed-forward neural network uses the ReLU transform for the nonlinear transformation. In the decoder, the model also uses a standard attentional mechanism to compute the attentional weights between the input and its corresponding context for the decoding operation.

Fig. 1.

Fig. 1

Basic structure of transformer

Vision transformer

Vision transformer (ViT) is the application of Transformer models to computer vision, especially for image classification tasks. ViT transforms images into sequences by segmenting them into different paths and encodes and classifies them using standard Transformer models. Compared to traditional CNN models, ViT is based on a multi-head self-attention mechanism [9], which can adapt to inputs of different sizes and shapes, provides better flexibility, and allows migration learning after large-scale pre-training.

Self-attention

Self-attention is a unique mechanism for computing the interaction between two elements in a sequence. Given a sequence input, each element can be used simultaneously as a query, key, and value. The attention function can be described as mapping a query and a set of key-value pairs to an output, where the query, key, value, and production are vectors. The outcome is computed as a weighted sum of values, where the weights assigned to each value are calculated by the compatibility function of the query with the corresponding key [8]. These weights can be used in a weighted sum to add the encoded vector at that location.

Multi-head attention

Multi-head attention is a combination of multiple sets of self-attention mechanisms, each mapping between a pair of queries, keys, and values, thus allowing the model to simultaneously attend to different representation subspaces from other locations [8]. In this process, multiple attention mechanisms form a "head", each getting a separate set of queries, keys, and values and generating the corresponding output [10]. It captures multiple feature representations simultaneously and detects the relationship between different parts, thus developing more contextual relevance and significantly improving the model's ability to learn the original drawing.

Other techniques

Recent studies have found that multilayer perceptual network (MLP) models excel in solving image tasks without convolution or self-attention mechanisms. Such models learn representations only through basic linear algebra operations, which can be computed repeatedly for different spatial locations and feature channels. Despite the long-term dominance of CNNs and ViT, simple MLP models perform well on specific kidney image processing tasks. This demonstrates that MLPs can learn efficient representations, opening up new ideas for deep learning. A typical example is that Saikia et al. [11] proposed a model MLP-UNet based only on MLP architecture for glomerular segmentation tasks. The results show that MLP-UNet performs on PAS-stained whole kidney images comparable to the pre-trained model TransUNet but with a 20% reduction in the number of parameters without needing pre-training. The research advancement of MLP models has proposed many novel architectures such as gMLP [12], ResMLP [13], ASMLP [14], Cyclemlp [15], etc. Transformer, CNN, and MLP perform differently on different tasks, and there is yet to be a unified optimal structure for deep learning. This section will focus on applying Transformer and its variant architectures to the kidney image processing task to find a network architecture more suitable for a specific task and thus advance the field.

Application of transformer in renal image processing

Currently, the Transformer mechanism has more applications in renal image processing, mainly including image classification [16], tumor lesion segmentation [17], renal organ segmentation [18, 19], etc. In addition, Transformer can achieve prognostic assessment of renal diseases [20, 21], provide treatment plans [22], help doctors write pathology reports [23], construct electronic medical records [24], and so on.

Using the Transformer mechanism, the application that can be used in renal image processing can realize the fast and accurate automated analysis and processing of renal images, improve clinicians' efficiency and diagnosis level, and bring new opportunities and challenges for renal disease research and clinical treatment.

Transformer applied to kidney image segmentation

Renal cancer is now considered one of the most common malignant tumors in urology, leading to a large number of deaths every year [19]. In the past 30 years, the number of new cases of renal cancer in China has dramatically increased from 110,700 to 598,300 cases [25]. Traditionally, the lesion areas of renal cancer patients are mainly identified by clinicians' depiction, which relies heavily on the clinical experience of doctors and is very time-consuming and prone to erroneous judgment. Accurate measurements from medical images can help doctors make accurate diagnoses and provide timely treatment. Medical image segmentation aims at identifying tumors and depicting different sub-regions of an organ from the corresponding background by assigning labels of predefined categories to each pixel in a medical image, e.g., CT [5] MRI [6]. Therefore, the emergence of automatic medical image segmentation techniques is crucial to improve the accuracy and efficiency of clinical diagnosis.

CNN-based and U-Net-based [26] medical image segmentation algorithms have performed better in recent years. Still, based on the limitations of convolutional operations, they cannot capture long-range relationships. To solve this problem, there have been some research works applying network models based on Transformer with improvements to kidney image segmentation and have achieved good results. In renal image segmentation, more application scenarios are renal organ segmentation, renal lesion segmentation, and automatic target area outlining.

In this paper, the goodness of segmentation performance is usually expressed in terms of the following metrics. Dice similarity coefficient (DSC): measures the overlap between the segmentation result and the ground truth. Hausdorff distance (HD): computes the maximum distance between two sets, assessing differences between the predicted boundary and the ground truth boundary. IOU (Intersection over Union): calculates the ratio of the intersection area to the union area of the predicted region and ground truth, reflecting the degree of overlap. MIoU (mean IOU): represents the average IOU values of multiple samples, offering a comprehensive evaluation of model performance. F1 Score: considers both precision and recall, providing a balanced assessment of classification model performance. AUC (area under the ROC curve): reflects the overall performance of a classification model by measuring the relationship between true and false positive rates at different thresholds. Accuracy (ACC): indicates the model's overall classification performance. Sensitivity: measures the correct identification rate of positive cases. Specificity: measures the correct identification rate of negative cases.

Models for renal image processing based on transformers typically employ simple random rotation data augmentation for preprocessing, using cross-entropy as the loss function and optimization methods such as SGD and Adam. Regularization techniques include dropout and weight decay. Key hyperparameters encompass the learning rate (usually ranging from 1e−4 to 1e−5), batch size (4 to 16), and dropout rate (0.1 to 0.3)[27].

Multi-organ segmentation of the abdomen

Accurate kidney organ segmentation can provide clinicians with important information, and the task is often integrated with abdominal multi-organ segmentation. In the abdominal multi-organ segmentation task, the algorithm needs to segment all the organs in the abdomen at once. The synapse dataset (https: //doi.org/10.7303/syn3193805) is the most common publicly available dataset for abdominal multi-organ segmentation. Previous researchers usually use CNN for multi-organ segmentation [28]. The algorithm needs to consider global and local information to improve further the segmentation accuracy, which led to a combined Transformer–CNN model. According to the way of combining CNN and Transformer, hybrid Transformer model methods are usually classified into three categories (as shown in Fig. 2):

Fig. 2.

Fig. 2

ac Shows the transformer model-based encoder method, the method of using the transformer model between encoder and decoder, and the decoder based on the transformer model

①Methods based on the encoder of the Transformer model.

②Methods using the Transformer model between the encoder and decoder.

③Methods utilizing the decoder based on the Transformer model-based decoder methods.

The first class of methods aims at extracting higher-quality coded features by taking advantage of the Transformer model's strength in modeling remote dependencies. Therefore, some studies have directly adopted a Transformer as an encoder. TransUNet is a successful attempt to introduce a Transformer into medical image segmentation tasks [29]. This model employs a Transformer as an encoder, which combines the strengths of a Transformer and U-Net and can extract the global context from the labeled image chunks. At the same time, the Transformer helps to capture spatial relations over long-range distances. Compared with V-Net, AttnUNet, and ViT, TransUNet performs better on multi-organ and heart segmentation tasks. In this way, TransUNet can handle large image sizes without the memory constraints of traditional encoder–decoder models. Similarly, there are many models inspired by the U-shaped architecture described above. For example, Atek [30] et al. designed a two-scale encoder (Swin-Transformer) U-shaped architecture (SwinT-Unet) to integrate the Shift Window (Swin) Transformer module and the Transformer Interactive Fusion (TIF) module. Models incorporating hierarchical SwinT modules into the decoder include UNETR [31], Swin-Unet [32], TransClaw U-Net [33], MISSFormer [34], and others. In order to aggregate features from multiple scales of an image, many methods propose a Transformer model block based on parallel shift windows to improve SwinT. For, Feng et al. [35] proposed the ConvWin-UNet structure, which combines the ConvWin, Transformer, and UNet and utilizes the W-MSA (weighted multi-scale aggregation) mechanism and convolution operation to accelerate the convergence and enrich the information exchange between patches. Using convolutional window operations for each convolutional layer in the encoder and decoder, the model achieves an Average DSC of 79.39% and an HD of 21.39 mm in the Synapse dataset. And to deal with multiple related tasks simultaneously, some methods U-Net based on the introduction of numerous parallel branches, such as Wang [36] designed a hybrid MT-UNet network; MTM first computes the self-factor efficiently by Local–Global Gaussian Weighted Self-Attention (LGG-SA) and then mines the interconnections between data samples by external attention (EA). The MT-UNet model achieved 78.59% and 90.43% DSC on the Synapse and ACDC datasets, respectively. Finally, a U-shaped model is constructed for accurate medical image segmentation. The method consistently outperforms Trans-Unet and other visual Transformers for complex-shaped organ segmentation (e.g., liver and left kidney).

Unlike the above approaches, the second class of techniques aims to enhance the network's modeling capability in cross-layer feature transfer by incorporating a Transformer between the encoder and decoder to improve the performance of the segmentation task. For example, Zhou et al. [37] proposed a new 3D Transformer model called nnFormer. nnFormer introduces a self-attention mechanism based on local and global volumes to learn 3D volume representations and uses skip attention instead of skip connections to improve further the performance, which can be used to use less computational cost to model global feature relationships efficiently. The model achieved 86.4%, 86.57%, and 92.06% DSC on the BraTs2016, 2017, Synapse, and ACDC datasets. Outperformed the LeViT-UNet-384s and TransUNet and was more advantageous in segmenting the pancreas and the stomach in terms of mean HD and DSC, respectively. Similarly, introducing the cross-attention cross-convolution Transformer module instead of skip connections are DAE-Former [38], DSTUNet [39], and so on.

The third class of methods incorporates the Transformer into the encoder. For example, the Trans-U model proposed by Guo et al. [40] uses the combined high-resolution positional data from CNN features and the global context stored by the Transformer to compensate for the loss of feature resolution caused by the Transformer. The DSC result of this model on the Synapse dataset is 76.56%, which is lower than the U-Net and attnUNet models. The main reason is that the Transformer cannot extract low-level morphological details in medical images. However, it works well in capturing high-level semantic information that helps classify. For this reason, it is proposed to combine the Transformer with U-Net and let the Transformer learn the positional features through the jump connection of U-Net so that the model can utilize the high-level semantics as well as consider the low-level morphology and may obtain better results in medical image segmentation.

Unlike the U-shaped model-based approach mentioned above, to enhance the Transformer network's ability in local feature extraction, Wang et al. [41] proposed the use of a pyramid structure to construct multiscale representations and deal with multiscale variations, firstly, using a lightweight convolutional layer to extract the low-level features and reduce the amount of data, and then, using the Transformer block and the convolution block's mixture of Transformer blocks and Convolutional blocks to handle high-level features. Models with similar ideas include ECT-NAS [42], C2Former [43], CASTformer [44], etc. Niu et al. [45] proposed a novel symmetric supervised network based on the traditional two-branch approach, which utilizes a symmetric supervisory mechanism to enhance the supervision of the network training and introduces a Transformer-based global feature alignment module to improve the global consistency between the two branches. Compared with the baseline SE-Net [46], the method improved by 16.9% and 25.98% on the MS-CMRSeg and CHAOS datasets, respectively, and showed significant performance in the multi-organ left and suitable kidney segmentation experiments with 78.46% and 81.45%, respectively. To solve the problem of information loss or resolution degradation due to downsampling or cutting of the input image by traditional Transformer, Themyr et al. [47] proposed a full-resolution memory (FINE) Transformer model, which learns the memory Token by learning the memory Token, which scales well in terms of memory and computational cost, and allows for localized image segmentation. It scales well and interacts with local image regions and all 3D volumetric regions. FINE has better performance and superiority over CNN and recent Transformer model baselines (e.g., CoTr [48] and nnFormer [37]) to focus entirely on high-resolution images. FINE obtained 87.1% DSC and better segmentation of small and complex organs such as the pancreas (Pa) and gallbladder (Gb).

Furthermore, to reduce the dependence on expensive labeled kidney data and to be more efficient in data acquisition, Wang et al. [49] proposed a cross-teaching semi-supervised medical image segmentation model based on CNN and Transformer models, aiming to improve the efficiency of automatic segmentation of multiple organs in abdominal CT. However, it was found in the validation on the FLARE2022 challenge dataset that the segmentation effect could have been more satisfactory. Although the separation network could segment most organs, the location of organs such as kidneys shifted. In contrast, Xin et al. [50] used U-Net, the backbone network of nnU-Net [51], as the final prediction network based on the combination of CNN and Transformer. An average DSC of 75.80% was obtained in the FLARE2022 challenge. To perform accurate organ segmentation without the need for manual annotation, Wang et al. [52] designed a self-supervised learning-based framework for one-time kidney organ segmentation, which is used to build a network model of global correlation between the reference samples (VALUE) and the desired segmentation samples (QUERY). Local features are extracted using a CNN, and then global features are removed from the local feature space via a Transformer. A semantic dependency embedding method introduces channel and spatial standard information into the Transformer to establish global corrections. The experiment compares the model with PANet [53], SENet [54] and SSL-ALPNet [55], and the test scenarios include observed organ settings (OO) and unobserved organ settings (UO). The results show that the model outperforms the others in the MICCAI 2015 CT and ISBI2019 MRI datasets. This demonstrates the effectiveness of using self-supervised learning to train Transformer and Convolutional Hybrid Networks to handle better OO and UO scenarios in medical image segmentation tasks.

Segmentation of renal tumors

Renal tumor segmentation refers to accurately segmenting the tumor region in the kidney from the surrounding normal tissue in medical images to quantitatively identify and locate the location and extent of the renal tumor, which can effectively assist doctors in tumor diagnosis, treatment, and monitoring.

To further improve the segmentation and save the running time and memory of the algorithm. Some approaches apply a Transformer to the encoder for feature extraction of kidney images. For example, Yu et al. [56] proposed UNesT, which employs a simplified and faster converging Transformer model encoder design to achieve local communication between positional information by hierarchically aggregating spatially adjacent patch sequences. The model performs state-of-the-art on the four datasets BTCV, KiTS2015, BraTS2021, and KiTS2021, outperforming the state-of-the-art integrated model SLANT [57] in a whole-brain segmentation task. Some methods learn more straightforward mappings, focusing on normalized pose and size images. For example, Barbera et al. [18] proposed a new CNN architecture that contains three consecutive modules: a regression module, a differentiable module, and a segmentation module. The architecture uses a spatial Transformer model network (STN) to normalize the input image to improve the accuracy of subsequent segmentation tasks. The differentiable module automatically localizes the regions of interest to reduce the manual labeling effort. The segmentation module uses a UNet-based architecture, and the model achieved good DSC scores (88.01% for kidneys and 87.12% for tumors) in the segmentation task for kidneys and tumors on pediatric data and KiTS19 data. Inspired by the hierarchical structure in the visual Transformer model, Yu et al. [58] proposed a method to segment kidney components using a 3D block aggregation Transformer model. They constructed a kidney substructure segmentation dataset containing 116 subjects. The model enables localized communication between sequential representations without changing the self-attention mechanism. It showed advanced performance in the segmentation task with a DSC metric of 84.67%. Boussaid et al. [59] used the spatial Transformer model and linear subspace projection to compare segmentation masks in feature space and to characterize global shape properties. The authors experimented on a 3D ultrasound dataset of left and right adult kidneys from 667 patients and obtained a DSC metric of 92.07%, demonstrating the validity and accuracy of the method.

Chen et al. [17] proposed a multi-stage 2.5D semantic segmentation network for multi-stage fine segmentation to address the high cost of computational resources for kidney mass segmentation. The first stage uses ResSENormUnet [60] combined with deep residual connectivity and attention mechanism to pre-segment the kidney and predict the approximate location and shape. In the second stage, fine segmentation is performed using the DenseTransUnet [61] network combined with dense connectivity and self-attention mechanism to more finely segment the contours of the kidney, tumor and cyst. Finally, post-processing operations based on 3D-connected regions remove possible false-positive segmentation results. The model obtained good DSC for kidney segmentation (Kidney: 94.3%, Tumor:77.79%, Cyst:70.99%), but the network approach can be improved for segmenting smaller kidneys, tumors, and cysts. To enhance the spatial modeling capability of the network while maintaining the efficient use of computational resources, Yang et al. [62] proposed that the EPT-Net network effectively combines the edge sensing and Transformer structures and introduces the Dual Position Transformer to enhance 3D spatial localization capability. Meanwhile, the Edge Weight Guidance module extracts edge information without additional network parameters. Good performance is demonstrated on the relabeled KiTS2019 dataset (KiTS19-M).

Outlining of the renal target area

Radiation therapy is one of the most crucial localized treatment modalities for abdominal malignancies (e.g., cervical, prostate, pancreatic, renal, and liver cancers). Depicting abdominal organs at risk (OARs) on CT images is essential during radiation therapy management [63]. The method currently used in clinical practice is manual contouring of CT images, which is often very tedious and time-consuming. The results also vary depending on the skill level of the observer, environment, or equipment type. Deep learning-based automated contouring techniques for segmenting OAR would help eliminate these problems and produce consistent results with minimal time and labor [64].

Traditionally, there are conditional generative adversarial network (GAN) techniques proposed by Seenia et al. [64] for semantic segmentation of OAR in CT images of organs such as kidneys and Pan et al. [65] for multi-organ segmentation of abdominal CT images utilizing a V-net-like structure, a U-shaped multilayer perceptron mixer (MLP-Mixer) and a convolutional neural network (CNN). These methods need to use the image feature information effectively. At the same time, Jiang et al. [66] proposed the MRRN-NBSA method incorporating self-attention to segment multiple key OARs of head and neck (HN) and abdominal organ (BTCV) datasets. Comparison of MRRN-NBSA with Unet using cross-attention (CCA), dual-SA, and transformer-based (UNETR) methods showed that MRRN-NBSA obtained a DSC of HN: 88% and BTCV: 86%. The technique applies NBSA in a decoder that incorporates interactions between regional contexts while extracting non-local attentional information in a fast and memory-efficient manner. Overall, the network extracts relevant feature sets to generate accurate segmentation of organs such as kidneys by combining a deep multiresolution residual network and nested block (SA) self-attention to take advantage of multiscale features and self-attention mechanisms. To address the limitations in global and local information feature fusion in the classical TransUnet model decoder, Jiang et al. [67] proposed BiFTransNet, which introduces the BiFusion module into the decoder stage to achieve effective global and local feature fusion by enabling feature integration from various modules. It is used in the Synapse dataset to develop automated gastrointestinal image segmentation to help radiation oncologists accurately target the X-ray beam to the tumor.

Summary of segmentation algorithms

A literature search reveals that TransUNet, Swin-Unet, AgDenseU-Net 2.5D, LeViT-UNet, ViTBI, UNETR, and HiFormer are the more popular algorithms in the field of renal medical image segmentation at present and show different degrees of advantages in kidney image segmentation tasks. We conducted a comprehensive evaluation of the above segmentation algorithms, as shown in Table 1.

Table 1.

Comparison of kidney image segmentation algorithm performance

Algorithms Datasets Evaluation indicators/results Main views and contributions Limitations
TransUNet [29] Synapse 2015/ACDC Synapse (DSC: 77.48%; Kidney (R): 81.87%; Kidney (L):77.02%; HD: 31.69 mm)/ACDC(DSC: 89.71%) TransUNet is the first successful attempt to introduce a Transformer into medical image segmentation. Combining CNN and Transformer in coding Transformer leads to a dramatic increase in the number of model parameters
IB-TransUNet [68] Synapse 2015

DSC: Kidney (R):79.87%

Kidney (L):83.89%

Using the UNet model to combine the information bottleneck (IB) with the Transformer More advantages in learning small organ features
Swin-Unet [32] Synapse 2015

DSC: 79.13%

HD: 21.55 mm

The information bottleneck block was innovatively introduced in the encoding; a hierarchical Swin Transformer model with moving windows is used as an encoder to extract contextual features. An asymmetric Swin Transformer model decoder with a patch extension layer is designed to perform the upsampling operation Higher dependency on large and diverse datasets with a large number of parameters and complexity
AgDenseU-Net 2.5D [60] KiTS 2021

DSC:

Kidney: 95%

Tumor: 87.8%

Cyst: 74.6%

Combining the features of AggRes (which enhances feature representation by aggregating residual connectivity and attention mechanisms) and DenseU-Net (which efficiently performs multi-scale feature fusion) Higher computation and memory consumption, longer training time
LeViT-UNet [69] Synapse/ACDC Synapse (DSC: 78.53%, Kidney (R): 80.25%, Kidney (L): 84.61%, HD: 16.84 mm)/ACDC (DSC: 90.32%) Using LeViT as the encoder of LeViT-UNet, combining LeViT Transformer with U-Net Some metrics do not reach SOTA, and the segmentation performance is imaged to some extent to reduce the computational complexity
ViTBIS [70] Synapse 2015 DSC: 80.45% Adding the Concat operator for merging features The dataset is more homogeneous, with fewer baselines for comparison

TransClaw

U-Net [33]

Synapse 2015 Synapse (DSC: 78.09%, HD: 26.38 mm)

Claw U-Net with Transformer

Combined/decoder dual-path design

Relatively homogenous data sets
After-Unet [71] Thorax-85/BCV/SegTHOR thorax

Thorax-85 (DSC: 92.32%)/BCV (DSC: 81.02%)/SegTHOR thorax

(DSC: 92.10%)

Both intra- and inter-slice long-distance cues were considered to guide segmentation Axis information is naturally provided mainly for 3D volume
TransBTSV2 [19]

KiTS 2019/

BraTS2019/

BraTS2020/

LiTS 2017

KiTS 2019 (DSC: KIdney: 97.37%, Tumor: 83.69%, Composite: 90.53%) Not limited to brain tumor segmentation (BTS) but focuses on general medical image segmentation, providing a powerful and efficient 3D baseline for the volumetric segmentation of medical images Mainly for 3D medical image segmentation tasks
UNETR [31] BTCV/MSD BTCV (AVG: 89.1%)/MSD (DSC: 71.1%, HD95: 8.822 mm) The Transformers encoder utilizes embedded 3D corpora to capture remote dependencies efficiently; the jump-join decoder combines extracted representations of different resolutions and predicts the segmentation output Mainly for 3D medical image segmentation
DBT-UNETR [72] BTCV AVG:80.3% An improved SwinUNETR is proposed based on UNETR with Swin Transformer as an alternative to Transformer No significant improvement in performance compared to UNETR
NnFormer [37]

Synapse 2015/

ACDC

Synapse (DSC: 87.40%)/ACDC(DSC: 91.78%) Utilizing a combination of cross-convolution and self-attention operations Little performance gain on the ACDC dataset
HiFormer [73] Synapse 2015 DSC:80.69% Two multi-scale representations were designed based on the Swin transformer module and CNN encoder, and the Double-Level Fusion (DLF) module was designed to finely fuse the global and local features of the two representations Single dataset
MPSHT [74]

Synapse 2015/

ACDC

Synapse (DSC: 79.76%, KIdney: 80.77%, HD: 21.55 mm)/ACDC(DSC: 91.80%) Based on the CNN-Transformer model hybrid model, to which the asymptotic sampling module is added Accuracy of segmentation to be improved
DSGA-Net [75]

Synapse 2015/

BraTs 2020/

ACDC

Synapse (DSC: 81.24%)/BraTs2020 (DSC: 85.82%)/ACDC(DSC: 91.34%) Add a Depth Separable Gating Visual Transformation (DSG-ViT) module to the code and propose a Hybrid Three-Branch Attention (MTA) module Considerable computational burden; consumes large amounts of GPU memory
MedNeXt [76] BTCV/AMOS22/KiTS19/BraTS21/AVG BTCV (DSC: 88.76%)/AMOS22 (DSC: 91.77%)/KiTS19 (DSC: 91.02%)/BraTS21 (DSC: 91.49%)/AVG (DSC: 88.01%) The use of ConvNeXt 3D and the extension of ConvNeXt blocks to upsampling and downsampling layers represents a modern deep architecture for medical image segmentation Deep Networks Dedicated to Medical Image Segmentation
MESTrans [77] COVID-DS36/GlaS/Synapse/I2CVB COVID-DS36 (DSC: 81.23%)/GlaS (DSC: 89.95%, IoU: 82.39)/Synapse (DSC: 77.48%, HD:31.69 mm)/I2CVB (DSC: 92.3%, IoU: 85.8) Propose a Multi-scale Embedding (MEB) and Multi-layer Spatial Attention Transformer structure (SATrans) to adjust the sensory field. Propose a Feature Fusion Module (FFM) for global learning between shallow and deep features The performance of small organ segmentation needs to be improved
ST-Unet [78] Synapse 2015/ISIC 2018 Synapse2015(DSC:78.86%, HD:20.37mm)/ISIC 2018(F1:90.94%, mIoU:85.26) Proposing a new Cross-Layer Feature Enhancement (CLFE) module for cross-layer feature learning with spatial and channel squeezing and excitation modules to highlight the saliency of specific regions The accuracy of segmentation needs to be improved
COTRNet [79] KiTS 2021

DSC:

Kidney:92.28%

Tumor:55.28%

Cyst:0.50.52%

Utilizing pre-trained ResNet to develop the encoder, in addition to adding deep supervised The accuracy of segmentation for masses and tumors needs to be improved
CS-Unet [80] Synapse 2015

DSC:82.21%

Kidney(R):79.52%

Kidney(L):85.28%

Design of convolutional Swin-Transformer (CST) module that merges convolution with multi-head self-attention and feed-forward networks Facing the challenge of dealing with long-range dependencies

Transformer applied to kidney image classification

Kidney image classification is categorizing kidney image data into different categories or labels. With deep learning technology, kidney images can be automatically analyzed and classified to provide more accurate and faster diagnostic results. This helps to improve the early detection and treatment of kidney diseases. Due to the complexity of morphological and structural features of kidneys and surrounding tissues, the task of renal image classification usually needs to consider different levels of features [81], including renal morphology, size, texture, and so on. Traditional CNN models have limitations in dealing with complex kidney morphological and structural features. In contrast, Transformer can extract multiple sets of feature representations in parallel and incorporate a fully connected layer to fuse and classify the features, thus improving the model performance [82]. Therefore, applying Transformer to the renal image classification task can improve the accuracy and sensitivity, especially for the classification of renal cysts, tumors, stones, etc., thus helping doctors to understand renal lesions more accurately and provide better treatment plans and prognosis assessment.

Classification network model based on the combination of Transformer and CNN networks

The first class of approaches applies the Transformer to an encoder–decoder structure, where the encoder–decoder consists of multiple identical layers, each containing an Attention mechanism and a feed-forward neural network. For example, the MT-ONet network [83], combines CNN, hybrid Transformer and LGG-SA into the encoder component of the proposed O-Net architecture to improve the classification accuracy. The second class of approaches uses the Attention mechanism between the encoder and decoder to capture the dependency between input and output. For example, the CTransPath [84] network uses a new Semi-Supervised Learning (SSL) strategy called Semantic Relatedness Contrastive Learning (SRCL), which utilizes the local features of CNNs mining capability and the global interaction capability of Transformer, which has some advantages in solving small sample data.

In diagnostic pathology, whole-slice images are typically huge and often have only overall labels and no labels corresponding to specific instances (e.g., cells or lesions). This leads to the fact that traditional supervised learning methods cannot be directly applied to this problem. To transform the weakly supervised classification problem into an overlooked learning problem, Shao et al. [16]. proposed a new framework called Multiple Instance Learning (MIL) to explore the correlation between different instances to solve the weakly supervised classification problem in pathological diagnosis based on the whole section images of the kidney, based on the MIL framework, the paper designs A Transformer model-based MIL (i.e., TransMIL), which can efficiently handle unbalanced/balanced and binary/multiple classification with good visualization and interpretability. TransMIL achieved an AUC of 93.09% and TCGA-NSCLC: 96.03% TCGA-RCC: 98.82% on the CAMELYON16 and TCGA datasets.

CNNs are more commonly used for renal image classification tasks than Transformer models; for example, Cicalese et al. [85] proposed an uncertainty-guided Bayesian Classification (UGBC) scheme for glomerular and renal level classification tasks. Qadir et al. [86] used a deep migration learning model based on the DenseNet201 network to classify the tumor, normal cysts and stone regions of the kidney. Aruna et al. [87] used networks such as CNN and VGG19 to diagnose polycystic kidneys, and the classification task covered cysts, tumors, and stones. Hossain et al. [88] used three classification methods, namely, EAnet, ResNet50, and a customized CNN model, to classify the four types in CT images of the kidney (cysts, normal, stones, tumors). Chanchal et al. [89] proposed the RCCGNet network for fully automated renal cell carcinoma grading from renal histopathology images.

Summary of classification algorithms

In kidney image classification, algorithms based on CNN or combining CNN and Transformer have become a hot research topic. These algorithms utilize the feature extraction capability of CNN and the sequence modeling capability of Transformer to improve the accuracy and efficiency of kidney image classification. In this paper, we summarize some crucial algorithms, including TransMIL, CTransPath and other algorithms and CNN and DNN-based algorithm models, and their performance is summarized and compared in detail in Table 2. This provides an opportunity to analyze their strengths and limitations in depth and provides a reference for future research and applications.

Table 2.

Comparison of kidney image classification algorithm performance

Algorithms Datasets Evaluation indicators/results Main views and contributions Limitations
TransMIL [16] CAMELYON16/TCGA-NSCLC/TCGA-RCC AUC: (CAMELYON16: 93.09%, TCGANSCLC: 96.03%, TCGA-RCC: 98.82%) Using multiple instance learning (MIL) to explore morphological and spatial information in images Mainly dealing with weakly supervised classification in whole-slice image (WSI)-based pathology diagnosis
CTransPath [84] TCGA-RCC AUC:99.1% Self-computation of localized window attention using Swin-Transformer as a backbone model Large amounts of unlabeled data are required
UGBC [85] private dataset ACC (glomerulus: 96.30%, Kidney: 96.60%) Assigning image labels based on kidney-level classification using a high-throughput batch labeling scheme to exploit label noise immunity associated with deep neural networks (DNNs) Dependence on the accuracy of label annotations
DenseNet201–Random Forest [86] CT KIDNEY DATASET: Normal-Cyst-Tumor and Stone ACC: 99.44% (cyst: 99.60%, kidney: 98.90%, tumor: 100%) Feature extraction using deep migration learning model DenseNet-201-Random Forest More resources are needed to train and use both models simultaneously
RCCGNet [89] KMC-kidney dataset/BreakHis dataset KMC-kidney (ACC: 90.14%, F1:89.06%)/BreakHis (ACC: 90.09%, F1: 88.90%) RCCGNet contains a shared channel residual (SCR) block, which shares information between two different layers and complements each other's shared data The model integration is complex

Multi-modal image alignment

Multimodal image alignment is aligning and matching renal image data from different modalities. By aligning images from other modalities, the correlations and implied relationships between them can be revealed, providing researchers with more information and insight. In clinical practice, doctors often need to refer to renal image data from multiple modalities simultaneously, such as MRI, CT, and ultrasound images. By aligning these images, the correlation analysis between different modalities can be realized, improving the accuracy of diagnosis and treatment decisions.

Chi et al. [90] proposed a new depth alignment pipeline for free-breathing 3D CT and 2D ultrasound (U/S) kidney scans. The pipeline consists of a feature network and a 3D–2D CNN-based alignment network. The feature network has hand-textured feature layers to reduce semantic gaps. The alignment network adopts the encoder–decoder structure of feature image mismatch (FIM), is first pre-trained with a retrospective dataset and training data generation strategy, i.e., the kidneys are uniformly aligned on the upper and lower axes on the CT images, and then the kidneys are aligned with the center of mass on the U/S images, and successfully achieves accurate alignment between kidneys on CT and U/S images. The pipeline solves the challenge of 3DCT–2DUS kidney alignment during free-breathing with a new network structure and training strategy and obtains a DSC of 96.88% and 96.39% in CT and U/S images, respectively.

Other clinical applications for transformer

In addition to intelligent analysis and intelligent diagnosis of medical images, the Transformer mechanism can also be applied to renal image detection, disease prediction, image alignment, electronic reports related to renal diseases, clinical decision models, etc. [91]. These renal image processing tasks involve large and complex image data, and the models constructed by traditional convolutional neural networks can hardly meet the actual clinical needs. Using an improved Transformer for kidney image data application is an efficient strategy that can help the medical imaging field accomplish quantitative analysis and clinical diagnosis of kidney images more accurately [92].

Transformer application for kidney disease prediction

The main clinical applications of renal ultrasonography include ruling out reversible causes of acute kidney injury, such as urinary tract obstruction, or identifying irreversible CKD to rule out unnecessary tests, such as renal biopsy [93]. Traditional methods of assessing kidney injury have relied on metrics such as kidney length, volume, cortical thickness, and echogenicity [94]. However, in recent years, advances in deep learning and computer vision have enabled machine learning and artificial intelligence techniques to more accurately and objectively assess kidney images, providing more comprehensive information to diagnose kidney injury and treatment decisions. Compared to traditional qualitative or semi-quantitative assessment methods, these techniques can reduce the influence of operator experience and subjective factors and provide more accurate assessment results [95].

Ma et al. [96] used a novel multimodal data model combining Transformer's bi-directional encoder representation and optical gradient boosters to improve CKD prediction. The MD-BERT-LGBM model was used in a CKD prediction experiment using over 3 /ls of medical data from 3295 participants and compared with traditional LR, LGBM and multimodal disease risk prediction algorithms. The results showed that MD-BERT-LGBM is expected to play an essential role in predicting and preventing CKD for clinical applications. Zeng et al. [97] constructed a sequential model for the prediction of acute kidney injury (AKI) induced by sepsis in the ICU. The attention-based sequential conduction model outperforms logistic regression, XGBoost, and RNN through a comprehensive performance evaluation. Its AUROC is 79.5% and AUPRC is 65.0%. Asif et al. [7] proposed a deep migration learning architecture based on the pre-trained VGG19 [98] model and Inception module, i.e., the architecture of the VGG19 model was customized by removing the fully connected layer and placing a randomly initialized plain Inception module and other coatings. It is used to detect major renal diseases from CT images. The experiments considered two migration learning approaches: feature extractor and fine-tuning. An AUC of 99.25% was achieved on 4000 renal CT images. The proposed model is of great benefit to urologists in detecting renal diseases. Shickelae et al. [99] designed a multi-stage end-stage renal disease (ESRD) prediction framework for ESRD based on the Transformer model. The framework was based on nonlinear dimensionality reduction, relative Euclidean pixel distance embedding, and spatial self-attention mechanisms for predictive modeling. Researchers developed a deep transformer network for coding WSI and predicting future ESRD using a dataset of 56 renal biopsy WSIs from patients with diabetic neuropathy at Seoul National University Hospital. The subjects had an AUC of 97% for the prediction of 2-year ESRD. Aboutalebi et al. [21] designed a clinician assessment-based dataset containing clinical and biochemical data of 1366 patients. Different machine learning models were developed and trained to predict kidney injury, including gradient-based augmented tree and deep Transformer architecture.

Transformer in electronic reporting

Electronic reporting has also been progressively applied in the medical field. Schuppe et al. [23] used the large-scale language Transformer model open source artificial intelligence ChatGPT, a patient diagnosed with bilateral renal cell carcinoma who underwent right partial and left total nephrectomy as well as episodic biliary atresia (BA) exhibited nephrotic syndrome (NS) signs and symptoms article reports were written. Yang et al. [24] described a methodology to develop a language model for reporting renal transplant pathology. The study aimed to answer two predefined questions: what rejection did the patient exhibit, and what was the grade of interstitial fibrosis and tubular atrophy (IFTA)? For this purpose, a corpus containing 3.4K renal transplant ports and 1.5 million words were used in the paper for pre-training in clinical BERT and fine-tuned with QA headers. Additionally, an extended renal BERT (i.e., exKidneyBERT) model was created, pre-trained and fine-tuned using the same corpus to capture the complex vocabulary of a narrow technical domain.

Application of transformer in decision-making systems

Zhang et al. [22] utilized the Decision Transformer model, an offline RL (reinforcement learning) paradigm for continuous time decision-making in the healthcare domain. In the paper, the model was generalized to a continuous-time decision-making scenario, considered past clinical measurements and treatments, and learned methods for suggesting future visit times and per-treatment schedules. Experimental results show that the continuous-time decision-making Transformer model can outperform its competitors. It has clinical utility in improving patients' health and prolonging their survival by learning high-performance strategies from log data generated using strategies of different quality levels.

Other applications summary

Kidney images play an essential role in clinical applications, and different algorithms have been proposed to achieve kidney image alignment and disease detection. Table 3 compares the performance and usage of several standard algorithms for clinical applications of kidney images.

Table 3.

Performance comparison of kidney image algorithms for other applications

Algorithms Datasets Evaluation indicators/results Main views and contributions Usage
VGG19 [87] Private dataset ACC: 98% VGG19 uses a deeper network structure and a small convolutional kernel for improved feature extraction Kidney cysts detection
EANet [88] CT KIDNEY DATASET: Normal-Cyst-Tumor and Stone ACC: 83.65% Introduction of attention mechanism, multi-scale feature fusion, efficient network design, cross-layer feature interaction Kidney cyst classification
ResNet50 [88] CT KIDNEY DATASET: Normal-Cyst-Tumor and Stone ACC: 87.92% Having introduced Residual Block and Batch Normalization Kidney cyst classification
MD-BERT-LGBM [96] private dataset

ACC: 78.12%

AUC: 85.15%

The model integrates a bi-directional encoder representation of the Transformer with an optical gradient lifter, a multimodal data model CKD disease prediction
KidneyRegNet [90] KiTS19/in-house datasets KiTS19 (DSC: 96.88%, Sensitivity: 0.9711, Specificity: 0.9667)/in-house (DSC: 96.39%, Sensitivity: 0.9736, Specificity: 0.9560) A new depth-alignment pipeline for free-breathing 3D CT and 2D U/S renal scans is proposed Kidney alignment
ChatGPT [23] NA NA The core algorithm is the Transformer, which combines the Transformer model's self-attention mechanism with the language model's generative power Nelson syndrome (NS) pathology report writing
MulGT [100] TCGA-KICA/TCGA-ESCA KICA (Typing: AUC: 98.44%, ACC: 93.89%, F1: 93.89%, Staging: AUC: 80.22%, ACC: 74.98%, F1: 72.55%)/ESCA (Typing: AUC: 97.49%, ACC: 92.81%, F1: 92.74%, Staging: AUC: 71.48%, ACC: 66.63%, F1: 65.73%) A domain knowledge-driven graph pooling module was designed to simulate diagnostic patterns for different analysis tasks WSI task diagnostics
Transformer [22] DIVAT (Database of Kidney Transplantation Medical Records) NA For use in medical fields where continuous-time decision-making is required Medical decision system
Transformer [99] Dataset of 56 renal biopsy WSIs in patients with DN AUC: 97% A multi-stage ESRD prediction framework based on the Transformer model For encoding WSI (whole-slice images) and predicting future ESRDs
Transformer [20] Private dataset

F1: 96.3%,

AUC: 98.9%

Predicting Kidney Transplant Function Using the Critical Mask Tensor of the Transformer Dot Product Attention Mechanism Predicting kidney transplant function
COVID-Net [21] Private dataset

Survival prediction: ACC:93.55%,

Kidney Injury Complications: ACC:88.05%

Proposing an interpretability-driven framework for building machine learning models to predict survival and kidney injury in patients with no coronary pneumonia from clinical and biochemical data Predicting survival and kidney injury in patients with new crown pneumonia
ExKidneyBERT [24] Private dataset

OneQA (ACC: 83.3%)

TwoQA (ACC: 95.8%)

Linguistic modeling of renal transplantation pathology reports Renal pathology reports

Discussion and outlook

This paper presents a comprehensive overview of Transformer model-based methods used for renal image processing tasks. After extensive comparisons and systematic analysis, compared with traditional CNNs, the Transformer model-based approach can capture the correlation between different locations in an image through the self-attention mechanism. It can consider global and local contextual information, improving the model's ability to understand and judge images. It shows excellent performance and potential to become the backbone network model in the renal disease image processing task.

In the clinic, the Transformer model-based approach can provide quantitative image analysis for doctors, thus assisting in the diagnosis and treatment planning of kidney disease. It has certain advantages in the segmentation and classification of kidney images: ① compared with other traditional models, the Transformer can effectively deal with long-range dependencies through the self-attention mechanism and can better capture the relationship between each part of the image, thus improving the accuracy of segmentation and classification; ② transformer model is more suitable for dealing with long sequence data and global information. The self-attention mechanism in Transformer allows interaction between arbitrary positional information without limiting parameter sharing and local sense fields, thus providing greater flexibility; ③ transformer model can be easily extended to handle multimodal data, such as the combination of image and text, which is advantageous in the task of multimodal information.

Although the Transformer model has unique advantages and potential in kidney image segmentation and classification tasks, some challenges and limitations must be addressed. For example, ① the Transformer may suffer from information loss when dealing with long-range dependencies compared to traditional CNNs;② the Transformer model consumes a large amount of computational resources, including memory and computational power, when dealing with large-scale image data. This may limit its feasibility and efficiency in practical clinical applications. ③ Transformer models usually require a large amount of training data for good generalization ability. However, in medical images, especially kidney images, acquiring large-scale labeled data is a challenging task.

Future research directions include the more effective integration of CNN and Transformer, the design of novel Transformer model architectures, the handling of multi-modal data, addressing unstructured data, and leveraging weak supervision and self-supervised learning to enhance the performance of clinical applications. The development of versatile and robust Transformer methods will facilitate improved analysis and application of clinical data. In the context of implementing this model in real-world medical diagnostics, three key challenges and considerations emerge. Firstly, privacy and security of data must be taken into account. Patient's private data should be appropriately handled and protected to prevent data leakage. Secondly, there is a need for diversity in training data. Currently, clinical sample sizes remain limited, resulting in constrained model generalization to different populations and disease types. Collaboration with more healthcare organizations is essential to collect large-scale clinical samples for model training to enhance its quality. It is worth noting that Transformer models typically require substantial training data to achieve strong generalization. However, obtaining extensive annotated data, especially in the field of medical imaging, such as kidney images, poses a challenging task. Lastly, it is necessary to validate the model's stability across multiple datasets, collecting diverse samples from different healthcare organizations for validation to test the model's robustness in various settings.

Through our review, we recognize the crucial importance of preprocessing methods in current kidney CT image processing. In our future work, we plan to further optimize and propose more effective CT image preprocessing approaches to overcome current challenges and limitations. Our method involves multi-step data preprocessing, including voxel size resampling, grayscale normalization, noise reduction, contrast enhancement, histogram equalization, region cropping, and data augmentation techniques. These comprehensive preprocessing steps aim to optimize model input, enhance performance, and improve generalization capabilities.

Conclusion

In kidney image analysis, diverse architectures and optimization techniques have significantly improved model performance. Transformer architectures for kidney image analysis are typically optimized in three main aspects: ① hybrid CNN and Transformer models, such as TransUnet [29]and U-Net variants, are employed to extract local features and learn global dependencies; ② introduction of 3D Transformer architecture, e.g., TransBTSV2 [19], focuses on learning CT/MRI 3D structural relationships, proving advantageous in volumetric image analysis compared to 2D models; ③transformer model modifications, including attention mechanism updates and depth increase for richer feature learning. For instance, the DSGA-Net [74] model introduces a Depth Separable Gated Visual Transformer (DSG-ViT) module to learn deeper features of kidney images. Multimodal data fusion, exemplified by MD-BERT-LGBM, combines different imaging modalities (CT, MRI, ultrasound) and text/label data, enhancing feature characterization. In summary, to enhance kidney image analysis task performance, appropriate model architectures need to be selected or modified based on data and task characteristics. We have summarized the features and performance of each model, providing a valuable reference resource for advancing and expanding kidney image analysis research.

Acknowledgements

Not applicable.

Abbreviations

CNN

Convolutional neural network

CKD

Chronic kidney disease

KSD

Kidney stones disease

CT

Computed tomography

MRI

Magnetic resonance imaging

RNNs

Recurrent neural networks

ViT

Vision Transformer

TIF

Transformer interactive fusion

W-MSA

Weighted multi-scale aggregation

LGG-SA

Local–global Gaussian-weighted self-attention

EA

External attention

FINE

Full-resolution memory

STN

Spatial Transformer model network

FCN

Full convolutional network

OARs

Organs at risk

GAN

Generative adversarial network

MLP-Mixer

U-shaped multilayer perceptron mixer

HN

Head and neck datasets

BTCV

Abdominal organ datasets

KiTS 2019

Kidney Tumor Segmentation Challenge 2019

KiTS 2021

Kidney Tumor Segmentation Challenge 2021

Synapse 2015

Synapse Multimodal MRI Segmentation and Classification Challenge 2015

ACDC

Automated Cardiac Diagnosis Challenge

Thorax-85

Thoracic Disease Screening in Chest Radiographs Dataset and Challenge

BCV

Brain Cancer Vision

SegTHOR thorax

Segmentation of THoracic Organs at Risk

BraTS2019

Multimodal Brain Tumor Segmentation Challenge 2019

BraTS2020

Multimodal Brain Tumor Segmentation Challenge2020

LiTS 2017

Liver Tumor Segmentation Challenge 2017

MSD

Medical Segmentation Decathlon

AMOS22

AMOS Medical Image Analysis Challenge 2022

BraTS21

Multimodal Brain Tumor Segmentation Challenge 2021

COVID-DS36

COVID-19 Diagnosis using Chest X-ray Images Dataset and Challenge

GlaS

Glasgow Retinal Image Analysis Challenge: Image Registration

ISIC 2018

2018 International Skin Imaging Collaboration: Skin Lesion Analysis Towards Melanoma Detection Challenge

DIVAT

Database of Kidney Transplantation Medical Records

CAMELYON16

Camelyon16: A Benchmark for Fully Automatic Multi-Path Segmentation of Lymph Nodes

TCGA

The Cancer Genome Atlas

SSL

Semi-supervised learning

SRCL

Semantic relatedness contrastive learning

MIL

Multiple instance learning

UGBC

Uncertainty-guided Bayesian Classification

U/S

Ultrasound

FIM

Feature image mismatch

AKI

Acute kidney injury

ESRD

End-stage renal disease

BA

Biliary atresia

NS

Nephrotic syndrome

IFTA

Interstitial fibrosis and tubular atrophy

MS-CMRSeg

MICCAI 2019 Multi-sequence Cardiac MRI Segmentation Challenge

CHAOS

ISBI 2019 Combined Healthy Abdominal Organ Segmentation Challenge

Author contributions

YY wrote the original manuscript and the design of the tables and illustrations. ZT and HW collaborated on the design of the study and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Grants from the National Natural Science Foundation of China (Grant No. 82170728).

Availability of data and materials

Not applicable.

Declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no conflict of interests regarding the publication of this paper.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Zhixian Tang, Email: tangzx@sumhs.edu.cn.

Huachun Weng, Email: Wenghc@sumhs.edu.cn.

References

  • 1.Kovesdy CP. Epidemiology of chronic kidney disease: an update 2022. Kidney Int Supplements. 2022;12(1):7–11. doi: 10.1016/j.kisu.2021.11.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Sassanarakkit S, Hadpech S, Thongboonkerd V. Theranostic roles of machine learning in clinical management of kidney stone disease. Comput Struct Biotechnol J. 2023;21:260–266. doi: 10.1016/j.csbj.2022.12.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Lemberger U, Pjevac P, Hausmann B, Berry D, Moser D, Jahrreis V, Özsoy M, Shariat SF, Veser JJU. The microbiome of kidney stones and urine of patients with nephrolithiasis. Urolithiasis. 2023;51(1):27. doi: 10.1007/s00240-022-01403-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Huang J, Leung DK-W, Chan EO-T, Lok V, Leung S, Wong I, Lao X-Q, Zheng Z-J, Chiu PK-F, Ng C-F. A global trend analysis of kidney cancer incidence and mortality and their associations with smoking, alcohol consumption, and metabolic syndrome. Eur Urol Focus. 2022;8(1):200–209. doi: 10.1016/j.euf.2020.12.020. [DOI] [PubMed] [Google Scholar]
  • 5.Nguyen H-G, Fouard C, Troccaz J. Segmentation, separation and pose estimation of prostate brachytherapy seeds in CT images. IEEE Trans Biomed Eng. 2015;62(8):2012–2024. doi: 10.1109/TBME.2015.2409304. [DOI] [PubMed] [Google Scholar]
  • 6.Huo Y, Liu J, Xu Z, Harrigan RL, Assad A, Abramson RG, Landman BA. Robust multicontrast MRI spleen segmentation for splenomegaly using multi-atlas segmentation. IEEE Trans Biomed Eng. 2017;65(2):336–343. doi: 10.1109/TBME.2017.2764752. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Asif S, Wenhui Y, Jinhai S, Ain QU, Yueyang Y, Jin H. Modeling a fine-tuned deep convolutional neural network for diagnosis of kidney diseases from CT images. In 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE; 2022:2571–2576.
  • 8.Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. Adv Neural Inf Process Syst. 2017, 30.
  • 9.Naseer MM, Ranasinghe K, Khan SH, Hayat M, Shahbaz Khan F, Yang M-H. Intriguing properties of vision transformers. Adv Neural Inf Process Syst. 2021;34:23296–23308. [Google Scholar]
  • 10.Li F, Lu X, Yuan J. Mha-corocapsule: multi-head attention routing-based capsule network for covid-19 chest x-ray image classification. IEEE Trans Med Imaging. 2021;41(5):1208–1218. doi: 10.1109/TMI.2021.3134270. [DOI] [PubMed] [Google Scholar]
  • 11.Saikia FN, Iwahori Y, Suzuki T, Bhuyan MK, Wang A, Kijsirikul B. MLP-UNet: Glomerulus Segmentation. IEEE Access 2023:1–1.
  • 12.Liu H, Dai Z, So D, Le QV. Pay attention to mlps. Adv Neural Inf Process Syst. 2021;34:9204–9215. [Google Scholar]
  • 13.Touvron H, Bojanowski P, Caron M, Cord M, El-Nouby A, Grave E, Izacard G, Joulin A, Synnaeve G, Verbeek J. Resmlp: feedforward networks for image classification with data-efficient training. IEEE Trans Pattern Anal Mach Intell. 2022;45(4):5314–5321. doi: 10.1109/TPAMI.2022.3206148. [DOI] [PubMed] [Google Scholar]
  • 14.Lian D, Yu Z, Sun X, Gao S. As-mlp: an axial shifted mlp architecture for vision. arXiv preprint, arXiv:2107.08391. 2021.
  • 15.Chen S, Xie E, Ge C, Chen R, Liang D, Luo PJ. Cyclemlp: A mlp-like architecture for dense prediction. arXiv preprint, arXiv:2107.10224. 2021. [DOI] [PubMed]
  • 16.Shao Z, Bian H, Chen Y, Wang Y, Zhang J, Ji X. Transmil: transformer based correlated multiple instance learning for whole slide image classification. Adv Neural Inf Process Syst. 2021;34:2136–2147. [Google Scholar]
  • 17.Chen Z, Liu H. 5 D Cascaded semantic segmentation for kidney tumor cyst. In International challenge on kidney and kidney tumor segmentation. Springer; 2021: 28–34.
  • 18.La Barbera G, Gori P, Boussaid H, Belucci B, Delmonte A, Goulin J, Sarnacki S, Rouet L, Bloch I. Automatic size and pose homogenization with Spatial Transformer Network to improve and accelerate pediatric segmentation. In 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI). IEEE; 2021:1773–1776
  • 19.Li J, Wang W, Chen C, Zhang T, Zha S, Wang J, Yu H. TransBTSV2: towards better and more efficient volumetric segmentation of medical images. arXiv preprint, arXiv:2201.12785. 2022.
  • 20.Milecki L, Kalogeiton V, Bodard S, Anglicheau D, Correas J-M, Timsit M-O, Vakalopoulou M. Contrastive masked transformers for forecasting renal transplant function. In International conference on medical image computing and computer-assisted intervention. Springer; 2022:244–254.
  • 21.Aboutalebi H, Pavlova M, Shafiee MJ, Florea A, Hryniowski A, Wong A. COVID-Net Biochem: an explainability-driven framework to building machine learning models for predicting survival and kidney injury of COVID-19 patients from clinical and biochemistry data. arXiv preprint, arXiv:2204.11210. 2022. [DOI] [PMC free article] [PubMed]
  • 22.Zhang Z, Mei H, Xu Y. Continuous-Time decision transformer for healthcare applications. In International conference on artificial intelligence and statistics. PMLR; 2023:6245–6262. [PMC free article] [PubMed]
  • 23.Schuppe K, Burke S, Cohoe B, Chang K, Lance RS, Mroch HJC. Atypical Nelson syndrome following right partial and left total nephrectomy with incidental bilateral total adrenalectomy of renal cell carcinoma: a chat generative pre-trained transformer (ChatGPT)-assisted case report and literature review. Cureus. 2023;15(3):e36042. doi: 10.7759/cureus.36042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Yang T. exKidneyBERT: a language model for kidney transplant pathology reports and the crucial role of extended vocabularies. Waterloo: University of Waterloo; 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Xu Q, Zhang T, Xia T, Jin B, Chen H, Yang X. Epidemiological trends of kidney cancer along with attributable risk factors in China from 1990 to 2019 and its projections until 2030: an analysis of the global burden of disease study 2019. Clin Epidemiol. 2023;15:421–433. doi: 10.2147/CLEP.S400646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In Medical image computing and computer-assisted intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer; 2015:234-241.
  • 27.De Raad K, van Garderen KA, Smits M, van der Voort SR, Incekara F, Oei E, Hirvasniemi J, Klein S, Starmans MP. The effect of preprocessing on convolutional neural networks for medical image segmentation. In 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI). IEEE; 2021:655–658
  • 28.Liu J, Sun H, Katto J. Learned image compression with mixed transformer-CNN architectures. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023:14388–14397.
  • 29.Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, Lu L, Yuille AL, Zhou Y. transunet: transformers make strong encoders for medical image segmentation. arXiv preprint, arXiv:2102.04306. 2021.
  • 30.Atek S, Mehidi I, Jabri D, Belkhiat DEC. SwinT-Unet: hybrid architecture for medical image segmentation based on Swin transformer block and Dual-Scale Information. In 2022 7th International conference on image and signal processing and their applications (ISPA); 8–9 May 2022. 2022:1–6.
  • 31.Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, Roth HR, Xu D. Unetr: Transformers for 3d medical image segmentation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision. 2022:574–584.
  • 32.Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M. Swin-unet: Unet-like pure transformer for medical image segmentation. In European conference on computer vision. Springer; 2022:205–218.
  • 33.Yao C, Hu M, Li Q, Zhai G, Zhang XP. Transclaw U-Net: Claw U-Net with transformers for medical image segmentation. In 2022 5th International Conference on Information Communication and Signal Processing (ICICSP); 26–28 Nov. 2022. 2022:280–284.
  • 34.Huang X, Deng Z, Li D, Yuan X. Missformer: an effective medical image segmentation transformer. arXiv preprint, arXiv:2109.07162. 2021. [DOI] [PubMed]
  • 35.Feng X, Wang T, Yang X, Zhang M, Guo W, Wang W. ConvWin-UNet: UNet-like hierarchical vision Transformer combined with convolution for medical image segmentation. Math Biosci Eng. 2023;20(1):128–144. doi: 10.3934/mbe.2023007. [DOI] [PubMed] [Google Scholar]
  • 36.Wang H, Xie S, Lin L, Iwamoto Y, Han X-H, Chen Y-W, Tong R. Mixed transformer u-net for medical image segmentation. In ICASSP 2022–2022 IEEE International conference on acoustics, speech and signal processing (ICASSP). IEEE; 2022:2390–2394.
  • 37.Zhou H-Y, Guo J, Zhang Y, Yu L, Wang L, Yu Y. nnformer: Interleaved transformer for volumetric segmentation. arXiv preprint, arXiv:2109.03201. 2021.
  • 38.Azad R, Arimond R, Aghdam EK, Kazerouni A, Merhof D. Dae-former: dual attention-guided efficient transformer for medical image segmentation. International Workshop on PRedictive Intelligence In MEdicine Cham: Springer Nature Switzerland 2022:83–95.
  • 39.Cai Z, Xin J, Shi P, Wu J, Zheng N. DSTUNet: UNet with efficient dense SWIN transformer pathway for medical image segmentation. In 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI); 28–31 March 2022. 2022:1–5.
  • 40.Guo S, Sheng S, Lai Z, Chen S. Trans-U: transformer enhanced U-Net for medical image segmentation. In 2022 3rd International conference on computer vision, image and deep learning & international conference on computer engineering and applications (CVIDL & ICCEA); 20–22 May 2022. 2022:628–631.
  • 41.Wang F, Wang B. Hybrid transformer and convolution for medical image segmentation. In 2022 International conference on image processing, computer vision and machine learning (ICICML); 28–30 Oct. 2022. 2022:156–159.
  • 42.Xu S, Quan H. ECT-NAS: searching efficient CNN-transformers architecture for medical image segmentation. In 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 9–12 Dec. 2021. 2021:1601–1604.
  • 43.Wang J, Zhao H, Liang W, Wang S, Zhang Y. Cross-convolutional transformer for automated multi-organs segmentation in a variety of medical images. Phys Med Biol. 2023;68(3):035008. doi: 10.1088/1361-6560/acb19a. [DOI] [PubMed] [Google Scholar]
  • 44.You C, Zhao R, Liu F, Dong S, Chinchali S, Topcu U, Staib L, Duncan J. Class-aware adversarial transformers for medical image segmentation. Adv Neural Inf Process Syst. 2022;35:29582–29596. [PMC free article] [PubMed] [Google Scholar]
  • 45.Niu Y, Luo Z, Lian S, Li L, Li S, Song H. symmetrical supervision with transformer for few-shot medical image segmentation. In 2022 IEEE International conference on bioinformatics and biomedicine (BIBM); 6–8 Dec. 2022. 2022:1683–1687.
  • 46.Roy AG, Siddiqui S, Pölsterl S, Navab N, Wachinger C. ‘Squeeze & excite’ guided few-shot segmentation of volumetric images. Med Image Anal. 2020;59:101587. doi: 10.1016/j.media.2019.101587. [DOI] [PubMed] [Google Scholar]
  • 47.Themyr L, Rambour C, Thome N, Collins T, Hostettler A. Memory transformers for full context and high-resolution 3D Medical Segmentation. In machine learning in medical imaging: 13th International Workshop, MLMI 2022, held in conjunction with MICCAI 2022, Singapore, September 18, 2022, Proceedings. Springer; 2022:121-130.
  • 48.Xie Y, Zhang J, Shen C, Xia Y. Cotr: efficiently bridging cnn and transformer for 3d medical image segmentation. In Medical image computing and computer assisted intervention–MICCAI 2021: 24th international conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part III 24. Springer; 2021:171-180
  • 49.Wang R. A semi-supervised multi-organ segmentation method via cross teaching between CNN and transformer. arXiv preprint, arXiv:2112.04894. 2022.
  • 50.Xin R, Wang L. Abdominal multi-organ segmentation using CNN and transformer. In MICCAI Challenge on fast and low-resource semi-supervised abdominal organ segmentation. Springer; 2022: 270–280.
  • 51.Isensee F, Jaeger PF, Kohl SA, Petersen J, Maier-Hein KH. nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods. 2021;18(2):203–211. doi: 10.1038/s41592-020-01008-z. [DOI] [PubMed] [Google Scholar]
  • 52.Wang B, Li Q, You Z. Self-supervised learning based transformer and convolution hybrid network for one-shot organ segmentation. Neurocomputing. 2023;527:1–12. doi: 10.1016/j.neucom.2022.12.028. [DOI] [Google Scholar]
  • 53.Wang K, Liew JH, Zou Y, Zhou D, Feng J. Panet: few-shot image semantic segmentation with prototype alignment. In proceedings of the IEEE/CVF international conference on computer vision. 2019:9197–9206.
  • 54.Qin R, Fu X, Lang P. PolSAR image classification based on low-frequency and contour subbands-driven polarimetric SENet. IEEE J Select Topics Appl Earth Observ Remote Sensing. 2020;13:4760–4773. doi: 10.1109/JSTARS.2020.3015520. [DOI] [Google Scholar]
  • 55.Ouyang C, Biffi C, Chen C, Kart T, Qiu H, Rueckert D. Self-supervised learning for few-shot medical image segmentation. IEEE Trans Med Imaging. 2022;41(7):1837–1848. doi: 10.1109/TMI.2022.3150682. [DOI] [PubMed] [Google Scholar]
  • 56.Yu X, Yang Q, Zhou Y, Cai LY, Gao R, Lee HH, Li T, Bao S, Xu Z, Lasko TA. Unest: local spatial representation learning with hierarchical transformer for efficient medical segmentation. Med Image Anal. 2022;90:102939. doi: 10.1016/j.media.2023.102939. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Huo Y, Xu Z, Xiong Y, Aboud K, Parvathaneni P, Bao S, Bermudez C, Resnick SM, Cutting LE, Landman BAJN. 3D whole brain segmentation using spatially localized atlas network tiles. Neuroimage. 2019;194:105–119. doi: 10.1016/j.neuroimage.2019.03.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Yu X, Tang Y, Zhou Y, Gao R, Yang Q, Lee HH, Li T, Bao S, Huo Y, Xu Z. Characterizing renal structures with 3D block aggregate Transformers. arXiv preprint, arXiv:2203.02430. 2022.
  • 59.Boussaid H, Rouet L. Shape feature loss for kidney segmentation in 3d ultrasound images. In. BMVC; 2021.
  • 60.Sun P, Mo Z, Hu F, Song X, Mo T, Yu B, Zhang Y, Chen Z. Segmentation of kidney mass using AgDenseU-Net 2.5 D model. Comput Biol Med. 2022;150:106223. doi: 10.1016/j.compbiomed.2022.106223. [DOI] [PubMed] [Google Scholar]
  • 61.Lin A, Chen B, Xu J, Zhang Z, Lu G, Zhang D. Ds-transunet: dual swin transformer u-net for medical image segmentation. IEEE Trans Instrum Meas. 2022;71:1–15. [Google Scholar]
  • 62.Yang J, Jiao L, Shang R, Liu X, Li R, Xu L. EPT-Net: edge perception transformer for 3D medical image segmentation. IEEE Trans Med Imaging. 2023;42:3229–3243. doi: 10.1109/TMI.2023.3278461. [DOI] [PubMed] [Google Scholar]
  • 63.Liao W, Luo X, He Y, Dong Y, Li C, Li K, Zhang S, Zhang S, Wang G, Xiao J. Comprehensive evaluation of a deep learning model for automatic organs at risk segmentation on heterogeneous computed tomography images for abdominal radiation therapy. Int J Radiat Oncol Biol Phys. 2023;117:994–1006. doi: 10.1016/j.ijrobp.2023.05.034. [DOI] [PubMed] [Google Scholar]
  • 64.Francis S, Jayaraj P, Pournami P, Puzhakkal N. ContourGAN: auto-contouring of organs at risk in abdomen computed tomography images using generative adversarial network. Int J Imaging Syst Technol. 2023;33:1494–1504. doi: 10.1002/ima.22901. [DOI] [Google Scholar]
  • 65.Pan S, Chang CW, Wang T, Wynne J, Hu M, Lei Y, Liu T, Patel P, Roper J, Yang X. Abdomen CT multi-organ segmentation using token-based MLP-Mixer. Med Phys. 2023;50(5):3027–3038. doi: 10.1002/mp.16135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Jiang J, Elguindi S, Berry SL, Onochie I, Cervino L, Deasy JO, Veeraraghavan H. Nested block self-attention multiple resolution residual network for multiorgan segmentation from CT. Med Phys. 2022;49(8):5244–5257. doi: 10.1002/mp.15765. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Jiang X, Ding Y, Liu M, Wang Y, Li Y, Wu Z. BiFTransNet: a unified and simultaneous segmentation network for gastrointestinal images of CT & MRI. Comput Biol Med. 2023;165:107326. doi: 10.1016/j.compbiomed.2023.107326. [DOI] [PubMed] [Google Scholar]
  • 68.Li G, Jin D, Yu Q, Qi M. IB-TransUNet: combining information bottleneck and transformer for medical image segmentation. J King Saud Univ Comput Inf Sci. 2023;35(3):249–258. [Google Scholar]
  • 69.Xu G, Zhang X, He X, Wu X. Levit-unet: make faster encoders with transformer for medical image segmentation. In Chinese conference on pattern recognition and computer vision (PRCV). Springer; 2023:42–53.
  • 70.Sagar A: Vitbis: Vision transformer for biomedical image segmentation. In MICCAI Workshop on distributed and collaborative learning. Springer; 2021:34–45.
  • 71.Yan X, Tang H, Sun S, Ma H, Kong D, Xie X. After-unet: axial fusion transformer unet for medical image segmentation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision. 2022:3971–3981.
  • 72.Tao H, Mao K, Zhao Y. DBT-UNETR: double branch transformer with cross fusion for 3D medical image segmentation. In 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 6–8 Dec. 2022. 2022:1213–1218.
  • 73.Heidari M, Kazerouni A, Soltany M, Azad R, Aghdam EK, Cohen-Adad J, Merhof D. Hiformer: hierarchical multi-scale representations using transformers for medical image segmentation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision. 2023:6202–6212.
  • 74.Zhao Y, Li J, Hua Z. MPSHT: multiple progressive sampling hybrid model multi-organ segmentation. IEEE J Transl Eng Health Med. 2022;10:1–9. doi: 10.1109/JTEHM.2022.3210047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Sun J, Zhao J, Wu X, Tang C, Wang S, Zhang Y. DSGA-Net: deeply separable gated transformer and attention strategy for medical image segmentation network. J King Saud Univ Comput Inf Sci. 2023;35(5):101553. doi: 10.1016/j.jksuci.2023.04.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Roy S, Koehler G, Ulrich C, Baumgartner M, Petersen J, Isensee F, Jaeger PF, Maier-Hein KH. Mednext: transformer-driven scaling of convnets for medical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2023:405–415.
  • 77.Liu Y, Zhu Y, Xin Y, Zhang Y, Yang D, Xu T. MESTrans: multi-scale embedding spatial transformer for medical image segmentation. Comput Methods Programs Biomed. 2023;233:107493. doi: 10.1016/j.cmpb.2023.107493. [DOI] [PubMed] [Google Scholar]
  • 78.Zhang J, Qin Q, Ye Q, Ruan T. ST-Unet: Swin Transformer boosted U-Net with cross-layer feature enhancement for medical image segmentation. Comput Biol Med. 2023;153:106516. doi: 10.1016/j.compbiomed.2022.106516. [DOI] [PubMed] [Google Scholar]
  • 79.Shen Z, Yang H, Zhang Z, Zheng S. Automated kidney tumor segmentation with convolution and transformer network. In International challenge on kidney and kidney tumor segmentation. Springer; 2021: 1–12.
  • 80.Liu Q, Kaul C, Wang J, Anagnostopoulos C, Murray-Smith R, Deligianni F. Optimizing vision transformers for medical image segmentation. In ICASSP 2023 - 2023 IEEE International conference on acoustics, speech and signal processing (ICASSP); 4–10 June 2023. 2023:1–5.
  • 81.Kalantar-Zadeh K, Jafar TH, Nitsch D, Neuen BL, Perkovic V. Chronic kidney disease. The lancet. 2021;398(10302):786–802. doi: 10.1016/S0140-6736(21)00519-5. [DOI] [PubMed] [Google Scholar]
  • 82.Sun L, Zhao G, Zheng Y, Wu Z. Sensing R: spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans Geosci Remote Sens. 2022;60:1–14. doi: 10.1109/TGRS.2022.3231215. [DOI] [Google Scholar]
  • 83.Zheng P. MT-ONet: mixed Transformer O-Net for Medical Image Segmentation. In 2022 International conference on sensing, measurement & data analytics in the era of artificial intelligence (ICSMD); 30 Nov.-2 Dec. 2022. 2022:1–4.
  • 84.Wang X, Yang S, Zhang J, Wang M, Zhang J, Yang W, Huang J, Han X. Transformer-based unsupervised contrastive learning for histopathological image classification. Med Image Anal. 2022;81:102559. doi: 10.1016/j.media.2022.102559. [DOI] [PubMed] [Google Scholar]
  • 85.Cicalese PA, Mobiny A, Shahmoradi Z, Yi X, Mohan C, Nguyen HV. Kidney level lupus nephritis classification using uncertainty guided Bayesian convolutional neural networks. IEEE J Biomed Health Inform. 2021;25(2):315–324. doi: 10.1109/JBHI.2020.3039162. [DOI] [PubMed] [Google Scholar]
  • 86.Qadir AM, Abd DF. Kidney diseases classification using hybrid transfer-learning densenet201-based and random forest classifier. Kurdistan J Appl Res. 2023:131–144.
  • 87.Aruna SK, Deepa N, Devi T. A deep learning approach based on CT images for an automatic detection of polycystic kidney disease. In 2023 International conference on computer communication and informatics (ICCCI); 23–25 Jan. 2023. 2023:1–5.
  • 88.Hossain MS, Hassan SMN, Al-Amin M, Rahaman MN, Hossain R, Hossain MI. Kidney disease detection from CT Images using a customized CNN model and deep learning. In 2023 International conference on advances in intelligent computing and applications (AICAPS); 1–3 Feb. 2023. 2023:1–6.
  • 89.Chanchal AK, Lal S, Kumar R, Kwak JT, Kini J. A novel dataset and efficient deep learning framework for automated grading of renal cell carcinoma from kidney histopathology images. Sci Rep. 2023;13(1):5728. doi: 10.1038/s41598-023-31275-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Yanling C, Yuyu X, Huiying L, Xiaoxiang W, Zhiqiang L, Jiawei M, Guibin X, Weimin H. KidneyRegNet: a Deep Learning Method for 3DCT-2DUS Kidney Registration during Breathing. arXiv preprint, arXiv:2305.13855. 2023. [DOI] [PMC free article] [PubMed]
  • 91.Han K, Wang Y, Chen H, Chen X, Guo J, Liu Z, Tang Y, Xiao A, Xu C, Xu Y. A survey on vision transformer. IEEE Trans Pattern Anal Mach Intell. 2022;45(1):87–110. doi: 10.1109/TPAMI.2022.3152247. [DOI] [PubMed] [Google Scholar]
  • 92.Yuan F, Zhang Z, Fang ZJPR. An effective CNN and Transformer complementary network for medical image segmentation. Pattern Recogn. 2023;136:109228. doi: 10.1016/j.patcog.2022.109228. [DOI] [Google Scholar]
  • 93.Kuo C-C, Chang C-M, Liu K-T, Lin W-K, Chiang H-Y, Chung C-W, Ho M-R, Sun P-R, Yang R-L, Chen K-T. Automation of the kidney function prediction and classification through ultrasound-based kidney imaging using deep learning. NPJ Digit Med. 2019;2(1):29. doi: 10.1038/s41746-019-0104-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Sawhney R, Malik A, Sharma S, Narayan V. A comparative assessment of artificial intelligence models used for early prediction and evaluation of chronic kidney disease. Decis Anal J. 2023;6:100169. doi: 10.1016/j.dajour.2023.100169. [DOI] [Google Scholar]
  • 95.Ma F, Sun T, Liu L, Jing H. Detection and diagnosis of chronic kidney disease using deep learning-based heterogeneous modified artificial neural network. Futur Gener Comput Syst. 2020;111:17–26. doi: 10.1016/j.future.2020.04.036. [DOI] [Google Scholar]
  • 96.Ma D, Li X, Mou S, Cheng Z, Yan X, Lu Y, Yan R, Cao S. Prediction of chronic kidney disease risk using multimodal data. In 2021 The 5th International conference on computer and data analysis. Sanya, China: Association for Computing Machinery; 2021: 20–25.
  • 97.Zeng G, Zhuang J, Huang H, Gao Y, Liu Y, Yu X. Continuous prediction of acute kidney injury from patients with sepsis in ICU settings: A sequential transduction model based on attention. In Proceedings of the 2022 International Conference on Intelligent Medicine and Health. Xiamen, China: Association for Computing Machinery; 2022: 31–37.
  • 98.Carvalho T, De Rezende ER, Alves MT, Balieiro FK, Sovat RB. Exposing computer generated images by eye’s region classification via transfer learning of VGG19 CNN. In 2017 16th IEEE international conference on machine learning and applications (ICMLA). IEEE; 2017:866-870
  • 99.Shickel B, Lucarelli N, Rao A, Yun D, Moon KC, Seok HS, Sarder P. Spatially aware transformer networks for contextual prediction of diabetic nephropathy progression from whole slide images. In Medical Imaging 2023: Digital and Computational Pathology. SPIE; 2023:129-140 [DOI] [PMC free article] [PubMed]
  • 100.Zhao W, Wang S, Yeung M, Niu T, Yu L. MulGT: multi-task graph-transformer with task-aware knowledge injection and domain knowledge-driven pooling for whole slide image analysis. arXiv preprint, arXiv:2302.10574. 2023.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from BioMedical Engineering OnLine are provided here courtesy of BMC

RESOURCES