Skip to main content
IEEE Journal of Translational Engineering in Health and Medicine logoLink to IEEE Journal of Translational Engineering in Health and Medicine
. 2022 May 23;10:1800508. doi: 10.1109/JTEHM.2022.3176737

Znet: Deep Learning Approach for 2D MRI Brain Tumor Segmentation

Mohammad Ashraf Ottom 1,2,, Hanif Abdul Rahman 2,3, Ivo D Dinov 2
PMCID: PMC9236306  PMID: 35774412

Abstract

Background: Detection and segmentation of brain tumors using MR images are challenging and valuable tasks in the medical field. Early diagnosing and localizing of brain tumors can save lives and provide timely options for physicians to select efficient treatment plans. Deep learning approaches have attracted researchers in medical imaging due to their capacity, performance, and potential to assist in accurate diagnosis, prognosis, and medical treatment technologies. Methods and procedures: This paper presents a novel framework for segmenting 2D brain tumors in MR images using deep neural networks (DNN) and utilizing data augmentation strategies. The proposed approach (Znet) is based on the idea of skip-connection, encoder-decoder architectures, and data amplification to propagate the intrinsic affinities of a relatively smaller number of expert delineated tumors, e.g., hundreds of patients of the low-grade glioma (LGG), to many thousands of synthetic cases. Results: Our experimental results showed high values of the mean dice similarity coefficient (dice = 0.96 during model training and dice = 0.92 for the independent testing dataset). Other evaluation measures were also relatively high, e.g., pixel accuracy = 0.996, F1 score = 0.81, and Matthews Correlation Coefficient, MCC = 0.81. The results and visualization of the DNN-derived tumor masks in the testing dataset showcase the ZNet model’s capability to localize and auto-segment brain tumors in MR images. This approach can further be generalized to 3D brain volumes, other pathologies, and a wide range of image modalities. Conclusion: We can confirm the ability of deep learning methods and the proposed Znet framework to detect and segment tumors in MR images. Furthermore, pixel accuracy evaluation may not be a suitable evaluation measure for semantic segmentation in case of class imbalance in MR images segmentation. This is because the dominant class in ground truth images is the background. Therefore, a high value of pixel accuracy can be misleading in some computer vision applications. On the other hand, alternative evaluation metrics, such as dice and IoU (Intersection over Union), are more factual for semantic segmentation. Clinical impact: Artificial intelligence (AI) applications in medicine are advancing swiftly, however, there is a lack of deployed techniques in clinical practice. This research demonstrates a practical example of AI applications in medical imaging, which can be deployed as a tool for auto-segmentation of tumors in MR images.

Keywords: Brain tumor, region segmentation, deep learning, augmentation, neural networks

I. Introduction

Recent advancements in brain imaging, information technologies, and digital health records have opened the door for enormous progress in holistic examination and deep phenotyping of the human brain. Structural, functional, diffusion, and spectroscopy explorations are now possible using Magnetic Resonance Imaging (MRI) [1], [2]. These advances are enabling health practitioners, radiologists, and medical scholars to peer into the anatomical and physiological organization of the brain, accurately diagnose normal and pathological development, maturation, and aging, as well as assist health professionals and clinicians. Manual inspection, segmentation, and analysis of MR images can be challenging for a number of reasons. The processing of large volumes of complex multiresolution data, including series of images, volumes, and participants; requires significant time, effort, and unique expert skills, e.g., long training process and collaborating teams of radiologists, neurologists, and data scientists. Detection of normal and pathological traits is subjective and includes a significant margin of error [1], [3]. Computational and Artificial Intelligence (AI) techniques [4][6] can significantly assist with building interpretable models, trained on existing medical images to automatically identify patterns, derive labels, and diagnose new cases [7][9]. In general, AI models are not intended to replace, but rather to complement, assist and enhance human medical expertise. Derived computational models can optimize and expedite scientific research and clinical practice. Medical image segmentation is the process of recognizing, detecting, and labeling anatomical parts, physiologically relevant areas, or the intrinsic network organization of the brain. Segmentation brain maps facilitate clinical decision-making and provide a core step in computer-aided diagnostics. The literature categorizes segmentation into three types: automatic, semi-automatic, and manual segmentation. Generally, there are three main models to localize brain region boundaries as curves in 2D images or surfaces in 3D volumes, e.g., brain tumor regions [10], [11]. These include traditional machine learning approaches, atlas registration, and deep learning methods. In classical machine learning image segmentation, expert-derived manual segmentation masks are employed to train a machine learning classifier. Label annotation tools are used to identify the target features, objects, or shapes. Machine learning algorithms are used for training and then for testing the model to detect analogous shapes and features in prospective out-of-bag (new) datasets. For instance, a statistical learning approach was used to model brain anatomical structure segmentation and develop a probabilistic brain atlas [12], [13]. Another example, OpenLabeling (python tool) is used to label tumors in medical images, and a support vector machine is used to produce the model. This approach is considered time-consuming and requires domain experts to obtain the initial manual annotations. In addition, the performance is relatively poor when the border between malignant and benign tissues is ambiguous [14].

Multi-Atlas segmentation (MAS) is a semi-automatic segmentation method that relies on prior atlases information. MAS was evolved from the single-atlas method but is considered insufficient for medical segmentation due to the lack of prior information, which leads to low performance and suboptimal accuracy [15]. As an alternative, multi-atlases with more prior information are used for segmenting novel images. For example, employing pairwise registration between each atlas image and the novel image, where the registration results are used to propagate the atlas labels into the space of the new image coordinates, which is followed by the selection of the most likely labels for each voxel [16].

In recent years, deep learning methods for medical image segmentation have received lots of attention, due to multiple reports with outstanding results for finding and predicting the target shapes in the images. The literature includes many algorithms for segmentation, but the fully convolutional network (FCN) and convolutional autoencoder methods are considered the most effective methods for medical image segmentation [17]. A fully conventional network (FCN) consists of a set of layers where each layer outputs is a three-dimensional array of size Inline graphic, where Inline graphic and Inline graphic are height and width respectively, and Inline graphic is the channel number. The input image is considered as the first layer, and the subsequent layers are the output of convolving previous layers, the basic elements of each layer are convolution, pooling, and activation functions. Writing Inline graphic for the data vector at location Inline graphic in a particular layer, and Inline graphicfor the following layer, these functions compute outputs Inline graphic by the following equation (eq. 1) [18]:

I.

where Inline graphic denotes the stride, Inline graphic is the kernel size, and Inline graphic determines the layer type. The final output layer in (FCN) will be the same size as the input layer (input image), However, the number of channels is determined by the classification classes number where one channel means one class which is normally used to obtain the predicted mask (desired output). Figure 1 shows the framework for the fully convolutional networks for medical image segmentation.

FIGURE 1.

FIGURE 1.

Fully convolutional network [18].

Autoencoder methods were utilized to extract features from images or inputs sample while attempting to keep most of the original information. The autoencoder consists of two main parts, the encoder and decoder. The encoder part is to encode the input images into smaller dimensional intermediate representations, where the decoder is to reconstruct the original input images from the intermediate representation. The loss function in autoencoders is calculated in terms of the similarity between input images and the reconstructed images [19], [20].

Unet is an extension of fully convolutional network architecture to work with a limited number of training data and to produce more accurate segmentation [21], Figure 2 shows the Unet architecture, which consists of a contracting path (left side) and an expansive path (right side). The left side of the network represents the normal architecture of a convolutional network, and each block contains two Inline graphic convolutions, rectified linear unit (ReLU), and a Inline graphic max pooling respectively; this process is called down-sampling. At the end of each contracting block, the image size is reduced, and feature channels are doubled. The right side of Unet architecture is called expansive path, which consists of an up-sampling of the feature map followed by a Inline graphic convolution that reduces the number of feature channels by half. Unet also applies concatenation between the output of each block on the left side of the diagram with the corresponding cropped block on the right side, cropping or matching arrays sizes are essential to combine arrays into one array, and then apply another two Inline graphic convolutions followed by a ReLU activation function. At the final block, a Inline graphic convolution is used to produce the output [22]. A key feature in the Unet architecture is the use of a large number of feature channels, and the use of skip connections. The purpose of skip connections in convolutional networks is to enhance gradient flow through a large number of layers to focus on more features in the image and to generate shapely segmentation [23].

FIGURE 2.

FIGURE 2.

Unet architecture [22], an extension and modification of fully convolutional network, receives a new MR image and produces the mask tracking the shape of the detected tumor.

In the current literature, some studies on MRI tumor segmentation use alternative methods, such as convolutional networks, random forests, and Unet. A hybrid approach of Unet and RESNET (residual networks) has been proposed by [24], [25] yields high performance on brain tumor segmentation, dice = 0.90. Ankari et al. [26] introduced genetic algorithms (GA) with convolutional networks to perform brain tumors segmentation. GA aims to obtain the optimal CNN structure rather than the standard trial way of building the convolutional network. This approach constructs several networks with the highest accuracy of 0.94 on the TCGA dataset. Several experiments were performed by [27] for brain tumor segmentation using Feature Pyramid Networks (FPNs), RESNET, and Unet on TCGA (The Cancer Genome Atlas [28]) dataset. The reported validation results show optimal performance with dice = 0.93 by using RESNET. A recent study [14] intended to build an effective segmentation system for brain tumors using deep learning methods, specifically Cascade Convolutional Neural Network (C-CNN). Their approach depends on a small amount of data rather than using the entire dataset, which overcomes high computation time constraints, restricts resource use, and reduces the overfitting problem. This method also reported competitive results with dice = 0.92. Another study [29]] modified the Unet framework by replacing the deconvolution block with the Nearest Neighbor method and two convolutional layers. It achieved promising results for segmenting brain tumors with dice = 0.88 and IoU = 0.86. Goetz et al. [30] used the ExtraTrees algorithm to segment brain tumors. Since ExtraTrees provides more randomness during the training phase and their reported dice = 0.83.

Automated clinical diagnosis, detection of clinical phenotypes, and forecasting of biomedical attributes and health phenotypes can be subjective and error-prone. AI methods can augment human experts by deriving unbiased, reliable (low variability) and quick computer models that can assist medical professionals with identifying patterns (such as tumors) and suggesting optimal interventions. The applications of AI in medicine are advancing swiftly. Yet, there is caution to expeditious, blind, and ubiquitous immersion of AI in all aspects of clinical practice. This research demonstrates one pragmatic case where the application of AI in medical imaging provides a robust estimation of the presence and dynamics of brain tumor development. Following broad and deep performance assessment, such AI applications can be deployed as desktop tools or mobile apps for automated detection, classification, and tracking of brain tumors using MR images.

II. Methodology

A. Dataset

The dataset used in this research study is a public benchmark dataset based on The Cancer Genome Atlas Low-Grade Glioma (TCGA - LGG) [31], [32]. The idea behind TCGA – LGG compilation is to build cancer images for research purposes and to study the relationship between phenotype to genotype in cancer and medical images research. The dataset was obtained during treatment and follow-up at multiple locations, as shown in Table 1 [31].

TABLE 1. Institutions and Individuals Contributed to TCGA-LGG Data Collection.

Institution Name Number of patients
Henry Ford Hospital, Detroit 46 patients
ThomasJefferson University, Philadelphia 16 patients
Case Western Reserve University, Cleveland 48 patients

Budaet et al. [32] investigated the dataset in [31] and made it available for computational research and medical image processing. In addition, our focus is on the FLAIR MRI images, which contain LGG enhanced tumor images. Neuroradiology experts manually reviewed, annotated, and labeled the FLAIR images for 110 patients. For pathological cases, the experts delineated the shape boundary of the FLAIR abnormality (tumor) for each image. The original MRI and the manually labeled masks represented the training dataset. The total number of image slices obtained was 3,929, with 2,556 labeled as normal and 1,373 abnormal (tumor). A few random samples of the dataset are shown in Figure 3.

FIGURE 3.

FIGURE 3.

(A) Samples of dataset images (B) the corresponding annotated ground truth (tumor mask).

B. Data Preprocessing

Randomly, the dataset was divided into training (3005 images), validation (393 images), and testing (432 images). The validation represents about 10% of the original dataset, testing images represent about 15% of the training dataset, and the remaining images are for training the model. Images resolution in the original dataset was Inline graphic, and due to extensive processing and computation time for the size of the original image, we resized data to Inline graphic pixels.

Imaging data augmentation became a popular pre-processing step in data science and AI applications, it is mainly used when training data is limited or when the amount of data is important to produce a better computer model, in addition, data augmentation can reduce the overfitting problem and increase model performance [33]. In this research, we used Albumentations, a free and open-source python library to perform the necessary images augmentation [34]. Albumentations processes provided a broad and holistic data augmentation. HorizontalFlip was used to produce new images from the existing images based on flipping the input around the y-axis of input horizontally. Albumentations VerticalFlip was utilized to generate new images by flipping the input around the x-axis vertically. RandomRotate90 was employed to rotate the input Randomly by 90 degrees. Transpose was applied to exchange rows and columns in the input image. Random affine transformations using ShiftScaleRotate were applied to generate new images based on shifting, scaling, and rotating the input images. Finally, all input images pixels were converted to floating-point elements in the range between 0 and 1. Figure 4 shows sample data augmented by Albumentations python library techniques.

FIGURE 4.

FIGURE 4.

Samples of data augmentation using albumentations python library techniques.

III. Znet Framework

In the proposed network architecture, we utilized the concepts of two well-known segmentation methods to construct a novel framework for MR images tumor segmentation; adversarial networks (AN) and Unet method. The framework uses the principle of skip connections and concatenating tensors from the AN and encoder-decoder design from the Unet method. Figure 5 shows the complete architecture of the proposed Znet method. The architecture consists of an encoding part (analytic downsampling) and a decoder part (synthetic upsampling).

FIGURE 5.

FIGURE 5.

Proposed architecture of the Znet.

The encoder part consists of five blocks, where each encoder block contains double convolutions combined by batch normalization and rectified activation function ReLU (equation 2) and followed by max-pooling. The output of the encoder block is concatenated with the encoder block input. Note, the encoder block input is interpolated to match the feature map of the encoder block output.

III.

where Inline graphic is the optional weight, Inline graphic and Inline graphic are the target and input respectively. The decoder part consists of five blocks, the same as the encoder block, except for the use of ConvTranspose2d instead of max-pooling.ConvTranspose2d is usually used to enlarge a tensor at the end of each decoder block, so we finally get the original image dimensions in the last decoder block. The final block is the output block which consists of a single convolution and sigmoid activation function (equation 3). We trained the model for 200 epochs using adaptive moment estimation (ADAM) optimizer [35] (equation 4),three channels Inline graphic pixels, a batch size of 32, and binary cross-entropy loss function [36]. Table 2 shows hyper-parameters used for training the proposed model, and Table 3 shows a sample of the architecture summary.

TABLE 2. Hyper-Parameters Used for Training the Proposed Model.

Parameter Name
Epoch 200
Optimizer Adam
Batch Size 32
Learning Rate Inline graphic
Input Size Inline graphic

TABLE 3. The Proposed Method Architecture Summary.

Layer (type) Output Shape Param #
Conv2d [−1, 64, 128, 128] 1,792
BatchNorm2d [−1, 64, 128, 128] 128
ReLU [−1, 64, 128, 128] 0
Conv2d [−1, 64, 128, 128] 36,928
BatchNorm2d [−1, 64, 128, 128] 128
ReLU [−1, 64, 128, 128] 0
: : :
: : :
ConvTranspose2d [−1, 128, 128, 128] 65,664
Conv2d [−1, 64, 128, 128] 73,792
BatchNorm2d [−1, 64, 128, 128] 128
ReLU [−1, 64, 128, 128] 0
Conv2d [−1, 64, 128, 128] 36,928
BatchNorm2d [−1, 64, 128, 128] 128
ReLU [−1, 64, 128, 128] 0
Conv2d [−1, 64, 128, 128] 12,352
BatchNorm2d [−1, 64, 128, 128] 128
ReLU [−1, 64, 128, 128] 0
OneConv [−1, 64, 128, 128] 0
Conv2d [−1, 1, 128, 128] 65
Sigmoid [−1, 1, 128, 128] 0
OutConv [−1, 1, 128, 128] 0

A. Evaluation Metrics

The most common metrics for evaluating the performance of medical MR images segmentation include pixel accuracy, mean intersection over union ( Inline graphic), and dice coefficient. However, pixel accuracy doesn’t perform well in case of class imbalance in images because the dominant class will overlook other classes and result in unexpected outcomes. On the other hand, dice and IoU are proven to be a better choices for semantic segmentation in general [37], [38]. IoU and Dice evaluation matrices are similar and positively correlated since DICE is twice the amount of IoU as shown in equations 5 and 6 to measure the similarity between ground truth A and predicted segmentation B. In our study, we also used pixel accuracy (eq. 7), the F1 score (eq. 8), and MCC (Matthews Correlation Coefficient) (eq. 9 for evaluating the proposed approach [39], (5)–(9), as shown at the bottom of the page.

graphic file with name M28.gif

B. Hardware Specification

The proposed framework ran on the server with the following hardware and software specifications: 2x 16-core Intel Xeon CPUs, 2x NVidia Titan 12GB GPUs, 128 GB RAM, 6 TB HDD storage, Ubuntu 18.04.5 LTS, Nvidia GPU driver v460.91, CUDA 11.2 + CuDNN 8.1, Torch v1.10.0, torchvision v0.11.1, and Spyder v4.2.5.

IV. Results

The training process on TCGA – LGG dataset for 200 epochs showed a validation dice of 0.96 for detecting and segmenting the brain tumor on MR images. Each epoch required 2-3 minutes using hardware specification in section 2.5. In data science applications, we usually divide the data into two parts - the training set, and the testing set. The training set is also split into two parts, the model training set and model validation set, the purpose of dividing the training set into two parts is to avoid overfitting during the training process, and to ensure that model evaluation (during training) is performed using an unseen set of data. Figure 6 shows the proposed algorithm training and validation performance (red and dark blue curves, respectively), and model loss is a light blue curve.

FIGURE 6.

FIGURE 6.

The proposed algorithm training and validation performance.

We compared the performance of the proposed framework with the well-known segmentation algorithm Unet. The testing was performed on 432 random MR images. The results show that Znet outperforms Unet in regards to dice, dice loss, pixel accuracy, F1 score, and MCC as shown in Table 4.

TABLE 4. Evaluation Metrics Comparison Between the Proposed Approach (Znet) and the Benchmark Algorithm Unet.

Unet Znet (proposed)
Testing DICE 0.8544 0.9158
Testing DICE_loss 0.1452 0.0839
Pixels Accuracy 0.9922 0.9955
F1 Score 0.7878 0.8098
MCC 0.7839 0.8089
Trainable parameters 14,788,929 44,384,833

Figure 7 shows samples of auto segmentation for MR images using the proposed framework and the Unet model. Part (A) shows the original images before the segmentation process, (B) shows the ground truth for tumor location and shape in the original image, (C) indicates the produced mask (location and shape) for the tumor using Znet model, (D) displays the constructed mask for tumor using Unet model, (E) shows the auto segmentation of tumor using the Znet model, (F) is the corresponding segmentation using Unet, and (G) is the overlap between Znet and Unet. The analysis of the 432 testing dataset indicates that the Znet model can predict and segment the tumor in MR images with a dice value of about 0.92 and dice loss of 0.08.

FIGURE 7.

FIGURE 7.

Visual results and comparison of MR images tumor segmentation using the proposed model (Znet) and the benchmark model Unet.

V. Discussion and Conclusion

In this work, we proposed a new approach for MR images segmentation based on the deep learning concept of convolutional network and data augmentation to utilize the available labeled images. The architecture relies on auto encoder-decoder, the concept of skip-connections, and residual neural networks, which requires combining the output of the previous layer with the next layer. Also features maps are required to map the dimensions between the input and the output of each layer. The benefit of skip-connections is to find alternative and further paths for the learning process and the gradient, which increase the probability of model convergence and avoid vanishing gradients dilemma [40]. The model was trained for 200 epochs using a server equipped with 2x NVidia Titan 12GB GPUs and showed a dice accuracy of 0.96 during training and about 0.92 for the testing dataset. Other measurements include pixel accuracy of 0.996, and 0.81for F1 score and MCC. For the benchmark Unet, the performance measures were 0.85 for dice, 0.992 for pixels accuracy, and 0.79 for F1 score and MCC. The evaluation metrics and the visualization of auto segmentation in the testing dataset (Figure 6) show the capability of the proposed approach for diagnosing and auto-segmentation of MR images of brain tumors. We can confirm that pixel accuracy is not a suitable evaluation measure for semantic segmentation in presence of class imbalance, just like in this case of MR images segmentation. Because the dominant class in ground truth images is the background. Also, high pixel accuracy can be misleading in some computer vision applications. On the other hand, some evaluation metrics, such as dice and IoU, provide more accurate assessment of the performance of alternative semantic segmentation techniques.

In the future, we will report about a prospective improvement and extension of the proposed Znet architecture to enable classification, extraction, parcellation, and prediction of the presence and extent of brain tumors using multichannel 3D MRI volumes. Predictive analytics and modeling of high-dimensional neuroimaging hyper-volumes is challenging for a number of reasons, e.g., complexities of representing higher-order data tensors, computational barriers, and the natural exponential increase of the state space of the optimization problem. Training supervised techniques such as Znet requires a large number of annotated ground truths (labels). Acquiring such large-scale testing and validation data is resource-intensive and time-consuming. We are also working on deep learning strategies to generate realistic high-dimensional and multimodal neuroimaging data along with ground truth labels. The transfer learning approach can be utilized along with the proposed Znet to optimize the training and minimize the costs of developing, validating, and embedding AI techniques in clinical practice settings.

Funding Statement

This work was supported in part by the Fulbright Award 2021/2022 (U.S. Department of State, ECA, JACEE, and IIE); in part by SOCR, University of Michigan Ann Arbor; in part by NSF under Grant 1916425, Grant 1734853, and Grant 1636840; and in part by NIH under Grant R01CA233487, Grant R01MH121079, Grant R01MH126137, and Grant T32GM141746.

References

  • [1].Mandal P. K., Mahajan R., and Dinov I. D., “Structural brain atlases: Design, rationale, and applications in normal and pathological cohorts,” J. Alzheimer’s Disease, vol. 31, no. s3, pp. S169–S188, Sep. 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Colby J. B., Soderberg L., Lebel C., Dinov I. D., Thompson P. M., and Sowell E. R., “Along-tract statistics allow for enhanced tractography analysis,” NeuroImage, vol. 59, no. 4, pp. 3227–3242, Feb. 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Tang Y.et al. , “The construction of a Chinese MRI brain atlas: A morphometric comparison study between Chinese and Caucasian cohorts,” NeuroImage, vol. 51, no. 1, pp. 33–41, May 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Ottom M. A., “Convolutional neural network for diagnosing skin cancer,” Int. J. Adv. Comput. Sci. Appl., vol. 10, no. 7, pp. 333–338, 2019. [Google Scholar]
  • [5].Dinov I. D., Data Science and Predictive Analytics. Cham, Switzerland: Springer, 2018. [Google Scholar]
  • [6].Dinov I. D. and Velev M. V., Data Science: Time Complexity, Inferential Uncertainty, and Spacekime Analytics. Berlin, Germany: Walter de Gruyter, 2021. [Google Scholar]
  • [7].Ahmad M. A. B., Mining Health Data for Breast Cancer Diagnosis Using Machine Learning. Canberra, ACT, Australia: Univ. of Canberra, 2013. [Google Scholar]
  • [8].Ottom M. A. and Nahar K. M., “Computer aided diagnosis for chronic kidney disease using data mining,” Int. J. Cyber-Secur. Digit. Forensics, vol. 8, no. 3, pp. 188–194, 2019. [Google Scholar]
  • [9].Al-Issa Y., Ottom M. A., and Tamrawi A., “eHealth cloud security challenges: A survey,” J. Healthcare Eng., vol. 2019, pp. 1–15, Sep. 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Kapur T., Egger J., Jayender J., Toews M., and Wells W. M., “Registration and segmentation for image-guided therapy,” in Intraoperative Imaging and Image-Guided Therapy. New York, NY, USA: Springer, 2014, pp. 79–91. [Google Scholar]
  • [11].Hiwaki O., “Novel technique for noninvasive detection of localized dynamic brain signals by using transcranial static magnetic fields,” IEEE J. Transl. Eng. Health Med., vol. 9, pp. 1–6, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Tu Z., Narr K. L., Dollár P., Dinov I., Thompson P. M., and Toga A. W., “Brain anatomical structure segmentation by hybrid discriminative/generative models,” IEEE Trans. Med. Imag., vol. 27, no. 4, pp. 495–508, Apr. 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Tang Y.et al. , “The construction of a brain atlas: A morphometric comparison study between cohorts,” NeuroImage, vol. 51, no. 1, pp. 33–41, May 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Ranjbarzadeh R., Kasgari A. B., Ghoushchi S. J., Anari S., Naseri M., and Bendechache M., “Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images,” Sci. Rep., vol. 11, no. 1, pp. 1–17, Dec. 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Doan N. T., de Xivry J. O., and Macq B., “Effect of inter-subject variation on the accuracy of atlas-based segmentation applied to human brain structures,” Proc. SPIE, vol. 7623, Mar. 2010, Art. no. 76231S. [Google Scholar]
  • [16].Iglesias J. E. and Sabuncu M. R., “Multi-atlas segmentation of biomedical images: A survey,” Med. Image Anal., vol. 24, no. 1, pp. 205–219, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Rehman M. U., Cho S., Kim J. H., and Chong K. T., “BU-Net: Brain tumor segmentation using modified U-Net architecture,” Electronics, vol. 9, no. 12, p. 2203, Dec. 2020. [Google Scholar]
  • [18].Long J., Shelhamer E., and Darrell T., “Fully convolutional networks for semantic segmentation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Boston, MA, USA, Jun. 2015, pp. 3431–3440. [DOI] [PubMed] [Google Scholar]
  • [19].Ghosh S., Das N., Das I., and Maulik U., “Understanding deep learning techniques for image segmentation,” ACM Comput. Surv., vol. 52, no. 4, pp. 1–35, Jul. 2020. [Google Scholar]
  • [20].Rashed-Al-Mahfuz M., Moni M. A., Uddin S., Alyami S. A., Summers M. A., and Eapen V., “A deep convolutional neural network method to detect seizures and characteristic frequencies using epileptic electroencephalogram (EEG) data,” IEEE J. Transl. Eng. Health Med., vol. 9, pp. 1–12, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Zhang J., “UNet-line by line explanation,” Towards Data Sci., Tech. Rep., 2019. [Online]. Available: https://towardsdatascience.com/
  • [22].Ronneberger O., Fischer P., and Brox T., “U-Net: Convolutional networks for biomedical image segmentation,” in Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Intervent. Munich, Germany: Springer, 2015, pp. 234–241. [Google Scholar]
  • [23].He K., Zhang X., Ren S., and Sun J., “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Las Vegas, NV, USA, Jun. 2016, pp. 770–778. [Google Scholar]
  • [24].Saha A., Zhang Y.-D., and Satapathy S. C., “Brain tumour segmentation with a muti-pathway ResNet based UNet,” J. Grid Comput., vol. 19, no. 4, pp. 1–10, Dec. 2021. [Google Scholar]
  • [25].Akella A.et al. , “Classifying multi-level stress responses from brain cortical EEG in nurses and non-health professionals using machine learning auto encoder,” IEEE J. Transl. Eng. Health Med., vol. 9, pp. 1–9, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [26].Anaraki A. K., Ayati M., and Kazemi F., “Magnetic resonance imaging-based brain tumor grades classification and grading via convolutional neural networks and genetic algorithms,” Biocybern. Biomed. Eng., vol. 39, no. 1, pp. 63–74, Jan./Mar. 2019. [Google Scholar]
  • [27].Ghosh S. and Santosh K. C., “Tumor segmentation in brain MRI: U-Nets versus feature pyramid network,” in Proc. IEEE 34th Int. Symp. Comput.-Based Med. Syst. (CBMS), Jun. 2021, pp. 31–36. [Google Scholar]
  • [28].Tomczak K., Czerwińska P., and Wiznerowicz M., “The cancer genome atlas (TCGA): An immeasurable source of knowledge,” Contemp. Oncol., vol. 19, no. 1A, p. A68, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [29].Kamrul Hasan S. M. and Linte C. A., “A modified U-Net convolutional network featuring a nearest-neighbor re-sampling-based elastic-transformation for brain tissue characterization and segmentation,” in Proc. IEEE Western New York Image Signal Process. Workshop (WNYISPW), New York, NY, USA, Oct. 2018, pp. 1–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [30].Goetz M., Weber C., Bloecher J., Stieltjes B., Meinzer H.-P., and Maier-Hein K., “Extremely randomized trees based brain tumor segmentation,” in Proc. BRATS Challenge-MICCAI, 2014, pp. 6–11. [Google Scholar]
  • [31].Clark K.et al. , “The cancer imaging archive (TCIA): maintaining and operating a public information repository,” J. Digit. Imag., vol. 26, no. 6, Dec. 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [32].Buda M., Saha A., and Mazurowski M. A., “Association of genomic subtypes of lower-grade gliomas with shape features automatically extracted by a deep learning algorithm,” Comput. Biol. Med., vol. 109, pp. 218–225, Jun. 2019. [DOI] [PubMed] [Google Scholar]
  • [33].Khalifa N. E., Loey M., and Mirjalili S., “A comprehensive survey of recent trends in deep learning for digital images augmentation,” Artif. Intell. Rev., vol. 55, pp. 2351–2377, Mar. 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [34].Buslaev A., Iglovikov V. I., Khvedchenya E., Parinov A., Druzhinin M., and Kalinin A. A., “Albumentations: Fast and flexible image augmentations,” Information, vol. 11, no. 2, p. 125, Feb. 2020. [Google Scholar]
  • [35].Kingma D. P. and Ba J., “Adam: A method for stochastic optimization,” 2014, arXiv:1412.6980.
  • [36].BCELoss PyTorch Documentation, Pytorch, 2019. [Online]. Available: https://pytorch.org/docs/stable/generated/torch.nn.BCELoss.html
  • [37].Tiu E., “Metrics to evaluate your semantic segmentation model,” Towards Datasci., 2019. [Online]. Available: https://towardsdatascience.com/metrics-to-evaluate-your-semantic-segmentation-model-6bcb99639aa2
  • [38].Jadon S., “A survey of loss functions for semantic segmentation,” in Proc. IEEE Conf. Comput. Intell. Bioinf. Comput. Biol. (CIBCB), Oct. 2020, pp. 1–7. [Google Scholar]
  • [39].Chicco D. and Jurman G., “The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation,” BMC Genomics, vol. 21, no. 1, pp. 1–13, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [40].Adaloglou N., “Intuitive explanation of skip connections in deep learning,” AI Summer, 2020. [Online]. Available: https://theaisummer.com/skip-connections/

Articles from IEEE Journal of Translational Engineering in Health and Medicine are provided here courtesy of Institute of Electrical and Electronics Engineers

RESOURCES