Abstract
Brain tumors are complex clinical lesions with diverse morphological characteristics, making accurate segmentation from MRI scans a challenging task. Manual segmentation by radiologists is time-consuming and susceptible to human error. Consequently, automated approaches are anticipated to accurately delineate tumor boundaries and quantify tumor burden, addressing these challenges efficiently. The presented work integrates a convolutional block attention module (CBAM) into a deep learning architecture to enhance the accuracy of MRI-based brain tumor segmentation. The deep learning network is built upon a VGG19-based U-Net model, augmented with depthwise and pointwise convolutions to improve feature extraction and processing efficiency during brain tumor segmentation. Furthermore, the proposed framework enhances segmentation precision while simultaneously incorporating tumor area measurement, making it a comprehensive tool for early-stage tumor analysis. Several qualitative assessments are used to assess the performance of the model in terms of tumor segmentation analysis. The qualitative metrics typically analyze the overlap between predicted tumor masks and ground truth annotations, providing information on the segmentation algorithms' accuracy and dependability. Following segmentation, a new approach is used to compute the extent of segmented tumor areas in MRI scans. This involves counting the number of pixels within the segmented tumor masks and multiplying by their area or volume. The computed tumor areas offer quantifiable data for future investigation and clinical interpretation. In general, the proposed methodology is projected to improve segmentation accuracy, efficiency, and clinical relevance compared to existing methods, resulting in better diagnosis, treatment planning, and monitoring of patients with brain tumors.
Keywords: area quantification, brain tumor analysis, lightweight U-Net model, magnetic resonance imaging, segmentation
1. Introduction
Brain tumor is a major health concern around the world, with a wide range of morphological traits and clinical manifestations [1]. The accurate segmentation of brain tumors using magnetic resonance imaging (MRI) scans is critical for diagnosis, therapy planning, and patient monitoring. Traditional segmentation approaches frequently rely on radiologists' manual delineation, which can be time-consuming, subjective, and susceptible to interobserver variability [2]. Deep learning techniques, notably convolutional neural network (CNN), have recently transformed medical image processing by allowing for the automated and exact segmentation of anatomical features and pathological lesions from MRI scans [3]. These works have demonstrated the usefulness and efficiency of CNN, in automated brain tumor segmentation from MRI scans. Moreover, it offers useful insights, methodology, and benchmark results for researchers looking to build similar approaches for brain tumor investigation.
Over the last decade, various segmentation algorithms have been presented for the detection of anomalies such as tumors and gliomas from brain MRI utilizing image segmentation. In 2014, a Gaussian mixture model–based segmentation approach was proposed by the authors [4] for brain MRI images. Recently, multiple studies have demonstrated the usefulness and efficiency of deep learning, particularly CNN-based approaches, for automated brain tumor segmentation from MRI scans [5]. These studies offer valuable insights, methodology, and benchmark results for researchers looking to build similar approaches for brain tumor investigation. Researchers have explored various modifications to CNN models, such as incorporating symmetric masks [6], leveraging ensemble architectures combining CNN and U-Net for BraTS2019 [7], and integrating SVM with CNN for BraTS-2018 [8]. Furthermore, a cascade deep learning model was proposed to determine tumor stages, assisting medical professionals [9]. A hybrid U-Net-based feature extraction model was introduced to find the complex anatomical structures and low discriminative features and applied on the BraTS and LGG-MRI datasets for the segmentation task [10]. Several performance metrics and experimental analyses were performed to validate the model. In 2023, a hybrid feature extraction model was designed by incorporating residual U-Net, attention, and squeeze–excitation modules to perform the segmentation [11], and for experimental analysis, publicly available BraTS-2020 dataset was employed. An enhanced U-Net architecture was presented in [12] that considered the inception block rather than the convolutional block during the skip connection phase utilizing multiple BraTS datasets for segmentation. Edge U-Net, constructed with an encoder-decoder framework from the U-Net architecture, was launched and customized for brain tumor segmentation using a T1-weighted MRI image dataset [13]. It integrates boundary-related information with basic MRI images to increase tumor localization accuracy. Segmentation and survival rate prediction were investigated in the same year, with a multimodal MRI BraTS dataset employed to examine the tumor for diagnosis and treatment planning [14]. In their work, they developed a hybrid deep learning architecture incorporating a convolutional normalized mean filter for preprocessing and a residual recurrent U-Net segmentation model. In [15], the authors developed a technique for brain tumor segmentation and survival prediction based on radiomics properties collected from segmented tumor regions. The work illustrates the effectiveness of deep learning–based segmentation approaches in conjunction with radiomics analysis for predicting patient outcomes and risk stratification in glioma patients. Havaei et al. [16] suggested a rapid and accurate technique for brain tumor segmentation that uses both local and broader contextual characteristics from the BraTS-2013 dataset. Naser and Deen [17] designed a segmentation framework based on U-Net and a pretrained VGG16 model to simultaneously segment, detect, and grade LGG tumors using the BraTS dataset. The Alzheimer's Disease Neuroimaging Initiative (ADNI) [18] and Open Access Series of Imaging Studies (OASIS) are other popular MRI datasets used for classification and segmentation specially for dementia cases [19–21]. The authors of [19] used the dataset and introduced multiple ensemble models in their study for characterizing and estimating Alzheimer's disease, a progressive neurological disorder. Ensemble deep learning approaches have shown promise in improving the accuracy and reliability of Alzheimer's disease diagnosis by integrating multiple data modalities and leveraging advanced machine learning techniques. In [20], the authors analyzed different models for this dataset that merged fuzzy logic with several deep learning and explainable artificial intelligence models. The objective behind these models was to handle more accurate and linguistically interpretable decision support systems for detection, segmentation, and classification.
According to the research, U-Net-based deep learning frameworks and the BraTS dataset are the most effective and commonly used for segmentation. Despite substantial advances in deep learning–based image segmentation, precisely defining brain tumors remains challenging, particularly in the face of diverse tumor morphology, fluctuating image contrast, and noise distortions. Furthermore, while segmentation gives useful information about tumor location and extent, quantifying the tumor burden using measures like tumor area is critical for determining disease progression, treatment response, and prognosis. This work presents accurate and efficient deep learning models for brain tumor segmentation from MRI scans, as well as methods for calculating tumor area. These methodologies are aimed at improving segmentation accuracy, efficiency, and clinical relevance compared to existing methods, resulting in better diagnosis, treatment planning, and monitoring of patients with brain tumors. In summary, possible contributions of the works are as follows:
• To design a convolutional block attention module (CBAM) with a VGG19-based U-Net architecture that has been augmented with depthwise and pointwise convolutions to boost feature extraction and processing efficiency during brain tumor segmentation.
• To develop an algorithm for calculating the area of segmented tumor regions in MRI scans, providing quantitative measurements for further analysis.
• Exploration of the clinical relevance of segmented tumor areas, highlighting their potential for aiding in diagnosis, treatment planning, and outcome prediction.
2. Theory and Method
MRI is a popular medical imaging technology that generates comprehensive images of soft tissues within the human body. Accurate segmentation of MRI images, especially for brain tumors, is crucial for diagnosis and treatment. Traditional segmentation approaches, such as thresholding, are limited in their capacity to handle the complexity and variability of MRI data. In contrast, deep learning techniques have demonstrated exceptional performance in medical image segmentation [22]. Among them, the U-Net architecture and its variations have grown in prominence due to their capacity to provide reliable segmentations despite low training data.
2.1. Related Works
Segmentation is the strategy of dividing an image into individual regions or entities. After effective image segmentation, objects can be identified based on their shape, texture, and color. The segmentation algorithm discovers abrupt changes in intensity, such as edges, and partitions regions based on predefined criteria.
2.1.1. Traditional Methods
Thresholding-based segmentation is most commonly used in classical image processing techniques. It relies on pixel intensity comparisons to differentiate foreground and background regions. These approaches are employed for image segmentation, with each pixel value being assessed according to the rule:
| (1) |
where T is the threshold value. This equation is widely incorporated into different segmentation algorithms. However, finding the optimal threshold value (T) is the most challenging task.
-
•Iterative thresholding: The iterative self-organizing data analysis technique algorithm (isodata) is one of the most prominent ways to find the optimal threshold value [23] in which the working structure of the model can be described as follows:
- 1. Compute μ1 and μ2 from the foreground and background pixels, respectively.
- 2. Tnew = (μ1 + μ2)/2.
- 3. Loop (until Tnew ≠ Told) do
- 4. Compute μ1 and μ2 from the foreground and background pixels, respectively, using new Tnew.
- 5. Told = Tnew,
- 6. Tnew = (μ1 + μ2)/2.
- 7. End Loop
T new is the final optimal threshold value.
• Ostu thresholding: The Otsu approach uses clustering to determine image thresholds, and it works with bimodal histograms [24] on the MRI for tumor segmentation. The approach essentially attempts to decrease within-class variance while maximizing between-class variance.
Overall, the model can be written as follows:
1. Create histogram from the gray scale image
2. Based on threshold value (T), separated the image into two classes: Class 1 (value < = T) and Class 2 (value > T)
3. Compute variance: σ2 = ∑(Xi − μ)2/N
4. Compute within class variance
2.1.2. Deep Learning–Based Approaches
U-Net is a standard deep learning model that was introduced in 2015, and it was introduced for biomedical image segmentation. Till now, it is the most widely used model due to its ability to achieve precise pixel-wise segmentation with limited annotated data [25]. The characteristic “U” shape of the architecture is created by the symmetric expanding and contracting channels. The fundamental purpose of this architecture is to solve the medical industry's lack of annotated data, and it was designed to consume less data while remaining fast and accurate.
The U-Net architecture comprises two channels: the encoder, which contracts, and the decoder, which expands. The encoder mixes convolutional and pooling layers to capture context while reducing spatial dimensions. The decoder expands the feature maps before combining them with the encoder's associated high-resolution features via skip connections. This structure enables the network to learn both global and local properties efficiently. The overall architecture is presented in Figure 1.
Figure 1.

Block diagram of U-Net model.
Several improvements of U-Net model has been done by integrating residual blocks from ResNet [26], lightweight features from EfficientNet [27], densely connected layers inspired by DenseNet [28]. Each of these U-Net versions expands on the basic architecture, addressing various challenges: the residual network solves the vanishing gradient problem and allows for the formation of deeper networks capable of learning complex characteristics, multiscale feature aggregation of EfficientNet layers improves the algorithm's efficiency and suitability for resource-constrained applications, and Dense U-Net effectively reuses features and creates a consistent gradient flow, reducing the likelihood of gradients disappearing. These developments, taken together, help to improve the reliability and precision of medical picture segmentation systems.
2.1.3. Hybrid Network–Based Approaches
Ensembling segmentation using CNN and U-Net is a powerful approach in medical imaging, particularly for tasks like brain tumor segmentation. This technique utilizes the capabilities of both designs to improve segmentation accuracy and reliability. Generally, CNN is used to extract features for identifying patterns, whereas the U-Net allows capturing spatial and contextual information to make precise pixel-level segmentation. Each model produces segmentation maps, which are then combined to yield the final prediction. This combination can be achieved using techniques such as weighted averaging or choosing the best-performing model for a specific subregion. The block diagram of the ensemble model for segmentation is presented in Figure 2.
Figure 2.

Block diagram of hybrid networks.
3. Proposed Segmentation and Area Quantification Method
The suggested segmentation algorithms, which are based on a modified VGG19-based U-Net architecture, constitute a powerful approach to MRI segmentation by combining the structural advantages of U-Net with the optimization benefits of attention learning. Its capacity to collect complex information, manage deep networks, and give accurate segmentation makes it an important tool in medical imaging, especially for difficult tasks like brain tumor segmentation. This architecture not only improves segmentation accuracy but also increases the clinical value of MRI for diagnosing and treating brain cancers. The architecture of the proposed algorithm is presented in Figure 3.
Figure 3.

Basic block diagram of the proposed model.
3.1. Preprocessing
Preprocessing utilizes intensity normalization, skull stripping, and image registration as critical for increasing the quality and consistency of MRI images prior to segmentation. To reduce the computational complexity of the model, the images are scaled to the range of [0, 1] using the formula below:
| (2) |
where I(x, y) represents the pixel value of the image. Data augmentation processes like rotation, scaling, and mirroring can strengthen deep learning models and increase their generalization performance. For augmentation, rotation and flipping are explored, with 45°, 90°, and 135° used to enhance the number of images in the training dataset.
3.2. Proposed Modified VGG19 With U-Net Model
This work proposes a deep learning architecture leveraging the VGG19-based U-Net model to efficiently and accurately segment brain tumors from MRI images. The model seeks to reduce computational complexity, improve segmentation accuracy, and shorten inference time while retaining a compact design with fewer parameters. The model keeps the fundamental ideas of the U-Net design, including an encoder–decoder structure with skip links, but makes many changes to improve efficiency. The encoder is made up of several convolutional blocks from the pretrained VGG19 network [29] that gradually downsample the input image while extracting high-level features. To prevent overfitting, a dropout layer is added after each convolutional block. The decoder mirrors the encoder but conducts upsampling rather than downsampling utilizing the CBAM [30]. The decoder model includes a CBAM employing depthwise and pointwise separable convolution layers, which improve feature extraction and processing efficiency during brain tumor segmentation. The architectural block diagram of the CBAM network is illustrated in Figure 4. Each upsampling procedure incorporates transposed convolutions, which improve the spatial resolution of the feature maps. Skip connections from the respective encoder layers are combined with the upsampled feature maps to preserve spatial information and ensure accurate localization.
Figure 4.

Proposed CBAM architecture.
The bottleneck layer works as a bridge between the encoder and decoder, holding the most significant filters and doing substantial feature extraction. The final layer is a 1 × 1 convolution, which lowers the number of channels to one and creates a binary segmentation mask that is the same size as the input image. A Sigmoid activation function is used to calculate pixel-wise probabilities for the existence of the tumor.
In this paper, CBAM processes are integrated into a U-Net architecture while keeping the model lightweight. The main advantage is that it can help focus on the most important spatial features during reconstruction, improving segmentation accuracy. These blocks are often integrated between the encoder's skip connections and the appropriate upsampled features in the decoder. The model's key advantage is the use of depthwise separable convolutions, which substantially reduce the amount of parameters and computational costs when compared to traditional convolutions, resulting in a lighter and faster model. Furthermore, skip connections assure that spatial resolution is maintained during upsampling, resulting in exact segmentation of tumor boundaries. The proposed modified VGG19-based U-Net model with CBAM architecture of the algorithm is presented as below Figure 5:
Figure 5.

Proposed modified VGG19-based U-Net with CBAM network.
3.3. Area Quantification
This study presents a segmentation model followed by tumor area calculation, which is essential for assessing tumor growth, treatment efficacy, and prognosis. Tumor area measurement plays a significant role in clinical decision-making, as it is closely associated with disease severity and patient outcomes. By integrating these measurements with clinical data, specialists can analyze their relevance for surgical and therapeutic planning. Additionally, tracking segmented tumor areas from multiple MRI scans allows medical professionals to monitor disease progression over time. Furthermore, tumor area quantification enables exploration of correlations with key prognostic indicators, such as tumor grade, patient survival, and therapy response. It is the process of converting a segmented region's pixel count into a physical area measurement based on pixel spacing information from MRI metadata. The flow diagram of area quantification is shown in Figure 6.
Figure 6.

Proposed area quantification flow diagram.
3.3.1. Pixel Counting
The number of pixels of a tumor from the segmented image can be calculated using the following formula:
| (3) |
where I indicate the segmented image with the desirable tumor sizes M and N. Here, I(i, j) = 1 denotes the pixel corresponding to the tumor region, while I(i, j) = 0 represents the remaining pixel values.
3.3.2. Physical Area Calculation
MRI metadata that includes pixel spacing values, which represent the physical dimensions of each pixel in the MRI image. Let Sx and Sy be the pixel spacing values in the x and y directions, respectively. The pixel's physical area is given by the following:
| (4) |
The physical area of the segmented tumor in square millimeters (mm2) is calculated by combining the pixel's physical area (PA) with the total number of pixels (NP).
| (5) |
4. Experimental Results
The suggested algorithm was validated for tumor segmentation and area quantification using brain MRI data. The key goals of this analysis were to evaluate the model's accuracy, efficiency, and segmentation quality, as well as to compare it to traditional U-Net designs.
4.1. Dataset
To conduct the investigation, publicly accessible BraTS datasets (Version 3) from the multimodal brain tumor segmentation challenge were used as a benchmark dataset for evaluating tumor segmentation algorithms [31–33]. The dataset represents 110 lower grade glioma (LGG) patients from the Cancer Genome Atlas collection that have sufficient fluid-attenuated inversion recovery (FLAIR) sequence and genomic cluster data available. It includes brain MR images and customized FLAIR abnormality segmentation masks taken from the Cancer Imaging Repository. These datasets contain MRI images of various tumor types, sizes, and locations, allowing for a full study of segmentation algorithms under different conditions. Figure 7 exhibits a few tumor and nontumor images from the dataset.
Figure 7.

Few sample images.
The BraTS collection, which includes multiparametric MRI images such as contrast-enhanced T1, T2, and FLAIR sequences, is primarily intended for brain tumor investigation. In contrast, datasets such as ADNI and OASIS focus on the analysis of neurodegenerative disorders like Alzheimer's [34]. Since this work primarily focuses on tumor segmentation and area quantification, the BRaTS dataset is considered here for the experimental evaluation and analysis.
4.2. Complexity Analysis
This work makes use of the Keras flops and parameter counts libraries for model complexity analysis. Table 1 displays the total number of trainable parameters (TPs), nontrainable parameters (NTPs), and floating point operations per second (FLOPs) for all models utilized here.
Table 1.
Total parameters (trainable and non-trainable) and FLOPs of all the models.
| Evaluation metrics | U-Net model | Residual U-Net model | VGG19 with U-Net | EfficientNet with U-Net | DenseNet with U-Net | Proposed model |
|---|---|---|---|---|---|---|
| TP | 31.0 × 106 | 41.0 × 106 | 31.2 × 106 | 24.0 × 106 | 24.9 × 106 | 20.4 × 106 |
| NTP | 5888 | 56,960 | 3840 | 153,872 | 225,216 | 2880 |
| FLOPs | 3.9 × 1012 | 1.3 × 1012 | 4.5 × 1012 | 0.5 × 1012 | 0.8 × 1012 | 2.1 × 1012 |
The investigation revealed that VGG19 and Residual U-Net have TP values of approximately 31 M and 41 M, respectively, which are higher than those of any other models. On the other hand, EfficientNet and DenseNet with U-Net have around 24 M parameters, making them more compact than the preceding two models. In contrast, the proposed model has the lowest TP (20.4 M), making it the lightest among all variants. The NTP values indicate that all other models utilize pretrained feature extractors with frozen layers. Although the proposed model is parameter-efficient, its computational cost in terms of FLOPs remains moderate. While it is not the lightest model overall, it achieves a balance between computational complexity and efficiency.
4.3. Preprocessing
The database contains 7858 brain MRI images, each with three RGB channels. Each image was scaled to 256 × 256 pixels using nearest neighbor interpolation and normalized to [0, 1]. The dataset is divided into three components: 70% for training, 15% for validation, and the remaining part for testing. Data augmentation was applied to increase the diversity of training images, incorporating random rotations (0.2), shifts (0.05), and zooms (0.05) with the nearest neighbor fill mode, which helps reduce overfitting. The model was trained for 150 epochs using the Adamax optimizer, a batch size of 40, and an initial learning rate of 0.00025. Additionally, a callback system was implemented to improve the model based on dice loss, ensuring better performance and stability during training. An early stopping mechanism with a tolerance of 15 epochs was adopted, and the best model was selected according to validation accuracy. To account for GPU memory limits, the model was trained employing variable precision.
4.4. Evaluation Metrics
The performance of segmentation algorithms is statistically assessed using measures such as accuracy, Dice coefficients (DSCs), and intersection over union (IoU).
1. Accuracy: Accuracy is the proportion of pixels properly categorized across all classes in an image. This metric measures how effectively it divides an image into discrete segments.
2. DSCs: DSC reveal the similarity of predicted segmentation with traditionally segmented output labels. It is similar to the F1 score and considers equally true and false positives, resulting in its more accommodating in specific circumstances.
3. IoU: IoU indicates the overlap between the predicted segmentation mask and the ground truth mask. The IoU value is computed by dividing the expected and actual masks' intersection area by their union area. This metric is used to evaluate the accuracy and efficacy of segmentation models in detecting and distinguishing objects within images.
where P and O are the estimated and manually segmented output label, respectively.
4.5. Segmentation Performance
Significant accuracy and efficiency are key in the field of segmentation, which also play important roles in algorithm evaluation. Although transformer-based segmentation models and self-supervised learning methods have demonstrated strong performance, this study primarily focuses on evaluating the effectiveness of U-Net-based approaches on the BRaTS dataset. Therefore, U-Net-based approaches were selected for comparison and evaluation. This paper presents a comprehensive analysis of several cutting-edge artificial neural network models, including U-Net, ResNet–U-Net, VGG19–U-Net, EfficientNet–U-Net, DenseNet–U-Net, and the proposed model. The performance of these models was assessed across training, validation, and testing datasets utilizing important metrics such as loss, accuracy, IoU, and DSC. Figure 8 depicts the training and validation curves, which provide critical insights into the model's learning process, ensuring that the model is both correct and generalizable.
Figure 8.

IoU, Dice coefficient, and loss curve of (a) U-Net-based, (b) ResNet with U-Net–based, (c) VGG19 with U-Net–based, (d) EfficientNet with U-Net model, (e) DenseNet with U-Net model, and (f) proposed VGG19 with U-Net–based.
The figure demonstrates that the suggested algorithm surpasses all standard models in terms of DSC, IoU, and loss. It achieves the highest segmentation accuracy with an IoU of 0.8940 and the lowest loss of 0.0561. The DenseNet with U-Net model algorithm follows closely with an IoU of 0.8864 and a loss of 0.0606. The EfficientNet with U-Net model also performs well, achieving an IoU of 0.8766 and a loss of 0.0661. The VGG19 with U-Net model shows significant improvement over the basic U-Net and Residual U-Net models, with an IoU of 0.8799 and a loss of 0.0643. The residual U-Net model has an IoU of 0.8147 and a loss of 0.1037, while the basic U-Net model has an IoU of 0.8204 and a loss of 0.0999. Furthermore, both the training and validation curves drop and plateau at identical places, indicating that the model is learning efficiently and without overfitting.
To effectively illustrate the superior outcome of the suggested brain tumor segmentation model, we organized the results for the training, validation, and testing datasets of different state-of-the-art approaches that are provided in Tables 2, 3, and 4.
Table 2.
Comparison of the proposed method with state-of-the-art methods (training).
| U-Net model | ResNet with U-Net | VGG19 with U-Net | EfficientNet with U-Net | DenseNet with U-Net | Proposed model | |
|---|---|---|---|---|---|---|
| Loss | 0.0999 | 0.1037 | 0.0643 | 0.0661 | 0.0606 | 0.0561 |
| Accuracy | 0.9981 | 0.9981 | 0.9987 | 0.9987 | 0.9987 | 0.9989 |
| IoU | 0.8204 | 0.8147 | 0.8799 | 0.8766 | 0.8864 | 0.8940 |
| DSC | 0.9001 | 0.8965 | 0.9358 | 0.9339 | 0.9395 | 0.9437 |
Table 3.
Comparison of the proposed method with state-of-the-art methods (validation).
| U-Net model | ResNet with U-Net | VGG19 with U-Net | EfficientNet with U-Net | DenseNet with U-Net | Proposed model | |
|---|---|---|---|---|---|---|
| Loss | 0.1265 | 0.1252 | 0.0928 | 0.1005 | 0.0975 | 0.0855 |
| Accuracy | 0.9975 | 0.9975 | 0.9985 | 0.9985 | 0.9981 | 0.9983 |
| IoU | 0.7780 | 0.7801 | 0.8320 | 0.8196 | 0.8236 | 0.8434 |
| DSC | 0.8735 | 0.8752 | 0.9073 | 0.8996 | 0.9023 | 0.9146 |
Table 4.
Comparison of the proposed method with state-of-the-art methods (test).
| U-Net model | ResNet with U-Net | VGG19 with U-Net | EfficientNet with U-Net | DenseNet with U-Net | Proposed model | |
|---|---|---|---|---|---|---|
| Loss | 0.1089 | 0.1560 | 0.0846 | 0.1161 | 0.0964 | 0.1066 |
| Accuracy | 0.9976 | 0.9973 | 0.9982 | 0.9978 | 0.9981 | 0.9984 |
| IoU | 0.8050 | 0.7345 | 0.8450 | 0.7935 | 0.8256 | 0.8088 |
| DSC | 0.8907 | 0.8440 | 0.9155 | 0.8837 | 0.9038 | 0.8935 |
Tables shows that the proposed algorithm consistently achieves the lowest loss values (0.0561 in training, 0.0855 in validation, and 0.1066 in testing), indicating it makes fewer errors compared to other models. It also shows the highest accuracy (0.9989 in training, 0.9983 in validation, and 0.9984 in testing), demonstrating its superior ability to make correct predictions. Additionally, the proposed model achieves the highest IoU (0.8940 in training, 0.8434 in validation, and 0.8088 in testing), showing it has the best overlap with the actual data. Finally, it achieves the highest DSC (0.9437 in training, 0.9146 in validation, and 0.8935 in testing), indicating it has the highest similarity to the ground truth. These metrics collectively highlight the proposed model's efficiency in learning and its robustness in segmentation tasks, demonstrating noticeable improvement over state-of-the-art methods. As demonstrated by the results, the improvement in training accuracy is marginal; however, the improvement in test results is significantly noticeable, exceeding 2%.
In a nutshell, the suggested modified VGG19- and CBAM-based U-Net model was effective in segmenting brain tumors from MRI images. The DSC and IoU scores were both very high, indicating accurate and consistent segmentation results. These results highlight the suggested model's stability and efficiency, making it a feasible option for advanced segmentation tasks.
4.6. Statistical Analysis
To comprehensively analyze the suggested strategy, this study uses statistical analysis, including standard deviation (SD), to provide more information about the model's performance and reliability. The statistical analysis (mean ± SD) is limited to the suggested model, as it is the primary contribution of this study, and assessing its consistency provides a better understanding of its reliability and effectiveness. The accuracy, IoU, and DSC of the proposed model were evaluated over seven trials in the statistical analysis conducted on training, validation, and test cases. The average scores are presented in Table 5. However, test cases are the primary focus of the analysis, as they provide the most reliable indication of the algorithm's performance. According to the results, the average accuracy, IoU, and DSC for the test cases are 0.9982 ± 0.000124, 0.8270 ± 0.0082, and 0.9046 ± 0.0050, respectively. These results suggest that the model's segmentation performance is both reliable and consistent.
Table 5.
Statistical analysis (mean ± SD) of the proposed algorithm.
| Accuracy | IoU | Dice | |
|---|---|---|---|
| Training | 0.9987 ± 0.000248 | 0.8785 ± 0.01952 | 0.9349 ± 0.0110 |
| Validation | 0.9984 ± 0.000026 | 0.8284 ± 0.0103 | 0.9054 ± 0.0063 |
| Test | 0.9982 ± 0.000124 | 0.8270 ± 0.0082 | 0.9046 ± 0.0050 |
4.7. Segmentation Results
Figure 9 depicts the visual representation of the suggested method. The figure shows that the suggested model accurately segmented the tumor from the MRI images. However, some tumors cannot be recognized using the suggested approach because the features of benign tumors are not well highlighted in MRI images.
Figure 9.

Segmentation result of the proposed algorithm.
Although the quantitative and qualitative study of the segmentation results reveals the model's success in certain circumstances, there are some specific characteristics of tumors where the model exhibits difficulty in segmenting as shown in Figure 10. Figure 10 demonstrates that the model confronts issues for intensity volatility, low contrast, and poorly defined boundaries of a tumor. Furthermore, the model has difficulties when the image contains much more background pixels than foreground pixels.
Figure 10.

Failure case from the segmentation results of the proposed algorithm.
4.8. Area Quantification
Accurate measurement of the tumor area is critical for clinical analysis and therapy planning. There is frequently an association between tumor size and cancer grade. Higher grade cancers grow faster and may cover larger areas. Accurate tumor area measurement can aid in grading the malignancy, which is critical for prognosis and therapy planning. The tumor area might also indicate the severity of the disease. Larger tumor regions are typically associated with a worse prognosis, as they may signify more advanced disease.
To find the tumor area, the segmented regions were transformed into tumor sizes using the pixel spacing information from the MRI scans. Each MRI slice was assigned a binary mask that indicated the existence of tumor regions, and the number of pixels corresponding to the tumor region in each mask was counted. The pixel count has been mapped to actual area using the pixel spacing values from the MRI images' DICOM metadata. The model consistently produces therapeutically meaningful values, as seen in Figure 11, which depicts the tumor area from a few segmented tumors.
Figure 11.

Area quantification from the segmented tumor in squared millimeters in a CBAM, spanning training, validation, and testing datasets. These findings demonstrate the model's noticeable capacity to effectively segment tumors, reduce mistakes, and remain very similar to ground truth annotations. The addition of tumor area assessment increases its clinical utility, giving precise and measurable data for diagnosis, treatment planning, and patient monitoring. Overall, the suggested model's efficiency, precision, and resilience make it an important tool for improving early-stage brain tumor diagnosis and therapeutic outcomes. Despite the positive results obtained by the provided model, its current capabilities are limited for certain clinical applications, such as reliably detecting cancer grade. Additional modifications are required to improve its effectiveness and dependability in such crucial medical situations.
5. Conclusion
This work offered a modified VGG19-based U-Net algorithm for brain tumor segmentation, followed by area quantification. Several investigations were accomplished to validate the proposed model, and the results were compared to standard state-of-the-art models regarding segmentation performance. The model obtained improved performance metrics by combining depthwise and pointwise separable convolutions.
Acknowledgments
The authors thank the Department of Computer Science and Engineering of Dhaka University of Engineering & Technology, Gazipur, for providing research support.
Data Availability Statement
Data sharing is not applicable to this article as no new data were created or analyzed in this study.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding
The authors thank the Department of Computer Science and Engineering of Dhaka University of Engineering & Technology, Gazipur, for providing research support.
References
- 1.Ashafuddula N. I. M., Islam R. ContourTL-Net: Contour-Based Transfer Learning Algorithm for Early-Stage Brain Tumor Detection. International Journal of Biomedical Imaging . 2024;2024:20. doi: 10.1155/2024/6347920.6347920 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Islam R., Imran S., Ashikuzzaman M., Khan M. M. A. Detection and Classification of Brain Tumor Based on Multilevel Segmentation With Convolutional Neural Network. Journal of Biomedical Science and Engineering . 2020;13(4):45–53. doi: 10.4236/jbise.2020.134004. [DOI] [Google Scholar]
- 3.Balwant M. K. A Review on Convolutional Neural Networks for Brain Tumor Segmentation: Methods, Datasets, Libraries, and Future Directions. Irbm . 2022;43(6):521–537. doi: 10.1016/j.irbm.2022.05.002. [DOI] [Google Scholar]
- 4.Balafar M. A. Gaussian Mixture Model Based Segmentation Methods for Brain MRI Images. Artificial Intelligence Review . 2014;41(3):429–439. doi: 10.1007/s10462-012-9317-3. [DOI] [Google Scholar]
- 5.Akkus Z., Galimzianova A., Hoogi A., Rubin D. L., Erickson B. J. Deep Learning for Brain MRI Segmentation: State of the Art and Future Directions. Journal of Digital Imaging . 2017;30(4):449–459. doi: 10.1007/s10278-017-9983-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Chen H., Qin Z., Ding Y., Tian L., Qin Z. Brain Tumor Segmentation With Deep Convolutional Symmetric Neural Network. Neurocomputing . 2020;392:305–313. doi: 10.1016/j.neucom.2019.01.111. [DOI] [Google Scholar]
- 7.Ali M., Gilani S. O., Waris A., Zafar K., Jamil M. Brain Tumour Image Segmentation Using Deep Networks. IEEE Access . 2020;8:153589–153598. doi: 10.1109/ACCESS.2020.3018160. [DOI] [Google Scholar]
- 8.Wu W., Li D., Du J., et al. An Intelligent Diagnosis Method of Brain MRI Tumor Segmentation Using Deep Convolutional Neural Network and SVM Algorithm. Computational and Mathematical Methods in Medicine . 2020;2020:10. doi: 10.1155/2020/6789306.6789306 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Ranjbarzadeh R., Bagherian Kasgari A., Jafarzadeh Ghoushchi S., Anari S., Naseri M., Bendechache M. Brain Tumor Segmentation Based on Deep Learning and an Attention Mechanism Using MRI Multi-Modalities Brain Images. Scientific Reports . 2021;11(1) doi: 10.1038/s41598-021-90428-8.10930 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Nizamani A. H., Chen Z., Nizamani A. A., Bhatti U. A. Advance Brain Tumor Segmentation Using Feature Fusion Methods With Deep U-Net Model With CNN for MRI Data. Journal of King Saud University-Computer and Information Sciences . 2023;35(9) doi: 10.1016/j.jksuci.2023.101793.101793 [DOI] [Google Scholar]
- 11.Rahim Khan W., Mustafa Madni T., Iqbal Janjua U., et al. A Hybrid Attention-Based Residual Unet for Semantic Segmentation of Brain Tumor. Computers, Materials and Continua . 2023;76(1):647–664. doi: 10.32604/cmc.2023.039188. [DOI] [Google Scholar]
- 12.Aboussaleh I., Riffi J., Mahraz A. M., Tairi H. Inception-UDet: An Improved U-Net Architecture for Brain Tumor Segmentation. Annals of Data Science . 2024;11(3):831–853. doi: 10.1007/s40745-023-00480-6. [DOI] [Google Scholar]
- 13.Allah A. M. G., Sarhan A. M., Elshennawy N. M. Edge U-Net: Brain Tumor Segmentation Using MRI Based on Deep U-Net Model With Boundary Information. Expert Systems with Applications . 2023;213 doi: 10.1016/j.eswa.2022.118833.118833 [DOI] [Google Scholar]
- 14.Zaitoon R., Syed H. Ru-Net2+: A Deep Learning Algorithm for Accurate Brain Tumor Segmentation and Survival Rate Prediction. IEEE Access . 2023;11:118105–118123. doi: 10.1109/ACCESS.2023.3325294. [DOI] [Google Scholar]
- 15.Isensee F., Kickingereder P., Wick W., Bendszus M., Maier-Hein K. H. Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Third International Workshop, BrainLes 2017, Held in Conjunction With MICCAI 2017 . Springer; 2017. Brain Tumor Segmentation and Radiomics Survival Prediction: Contribution to the Brats 2017 Challenge; pp. 287–297. [Google Scholar]
- 16.Havaei M., Davy A., Warde-Farley D., et al. Brain Tumor Segmentation With Deep Neural Networks. Medical Image Analysis . 2017;35:18–31. doi: 10.1016/j.media.2016.05.004. [DOI] [PubMed] [Google Scholar]
- 17.Naser M. A., Deen M. J. Brain Tumor Segmentation and Grading of Lower-Grade Glioma Using Deep Learning in MRI Images. Computers in Biology and Medicine . 2020;121 doi: 10.1016/j.compbiomed.2020.103758.103758 [DOI] [PubMed] [Google Scholar]
- 18.Wyman B. T., Harvey D. J., Crawford K., et al. Standardization of Analysis Sets for Reporting Results From ADNI MRI Data. Alzheimer’s & Dementia . 2013;9(3):332–337. doi: 10.1016/j.jalz.2012.06.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Tanveer M., Goel T., Sharma R., et al. Ensemble Deep Learning for Alzheimer’s Disease Characterization and Estimation. Nature Mental Health . 2024;2(6):655–667. doi: 10.1038/s44220-024-00237-x. [DOI] [Google Scholar]
- 20.Tanveer M., Sajid M., Akhtar M., et al. Fuzzy Deep Learning for the Diagnosis of Alzheimer’s Disease: Approaches and Challenges. IEEE Transactions on Fuzzy Systems . 2024;32(10):5477–5492. doi: 10.1109/TFUZZ.2024.3409412. [DOI] [Google Scholar]
- 21.Sharma R., Goel T., Tanveer M., Lin C.-T., Murugan R. Deep-Learning-Based Diagnosis and Prognosis of Alzheimer’s Disease: A Comprehensive Review. IEEE Transactions on Cognitive and Developmental Systems . 2023;15(3):1123–1138. doi: 10.1109/TCDS.2023.3254209. [DOI] [Google Scholar]
- 22.Gordillo N., Montseny E., Sobrevilla P. State of the Art Survey on MRI Brain Tumor Segmentation. Magnetic Resonance Imaging . 2013;31(8):1426–1438. doi: 10.1016/j.mri.2013.05.002. [DOI] [PubMed] [Google Scholar]
- 23.Velasco F. D. Thresholding Using the Isodata Clustering Algorithm. IEEE Transactions on Systems, Man, and Cybernetics . 1980;10(11):771–774. doi: 10.1109/TSMC.1980.4308400. [DOI] [Google Scholar]
- 24.Nyo M. T., Mebarek-Oudina F., Hlaing S. S., Khan N. A. Otsu’s Thresholding Technique for MRI Image Brain Tumor Segmentation. Multimedia Tools and Applications . 2022;81(30):43837–43849. doi: 10.1007/s11042-022-13215-1. [DOI] [Google Scholar]
- 25.Ronneberger O., Fischer P., Brox T. Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference . Springer; 2015. U-Net: Convolutional Networks for Biomedical Image Segmentation; pp. 234–241. [Google Scholar]
- 26.He K., Zhang X., Ren S., Sun J. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition . IEEE Conference; 2016. Deep Residual Learning for Image Recognition; pp. 770–778. [Google Scholar]
- 27.Tan M., Le Q. International Conference on Machine Learning . PMLR; 2019. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks; pp. 6105–6114. [Google Scholar]
- 28.Huang G., Liu Z., Van Der Maaten L., Weinberger K. Q. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition . IEEE Conference; 2017. Densely Connected Convolutional Networks; pp. 4700–4708. [Google Scholar]
- 29.Simonyan K., Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition. 2014. https://arxiv.org/abs/1409.1556 .
- 30.Woo S., Park J., Lee J.-Y., Kweon I. S. Proceedings of the European conference on computer vision (ECCV) Springer Conference; 2018. CBAM: Convolutional Block Attention Module; pp. 3–19. [Google Scholar]
- 31.Buda M., Saha A., Mazurowski M. A. Association of Genomic Subtypes of Lower-Grade Gliomas With Shape Features Automatically Extracted by a Deep Learning Algorithm. Computers in Biology and Medicine . 2019;109:218–225. doi: 10.1016/j.compbiomed.2019.05.002. [DOI] [PubMed] [Google Scholar]
- 32.Mazurowski M. A., Clark K., Czarnek N. M., Shamsesfandabadi P., Peters K. B., Saha A. Radiogenomics of Lower-Grade Glioma: Algorithmically-Assessed Tumor Shape Is Associated With Tumor Genomic Subtypes and Patient Outcomes in a Multi-Institutional Study With the Cancer Genome Atlas Data. Journal of Neuro-Oncology . 2017;133(1):27–35. doi: 10.1007/s11060-017-2420-1. [DOI] [PubMed] [Google Scholar]
- 33.Pedano N., Flanders A. E., Scarpace L., et al. The Cancer Genome Atlas Low Grade Glioma Collection (TCGA-LGG) (Version 3) 2016. [DOI]
- 34.Sajid M., Sharma R., Beheshti I., Tanveer M., for the Alzheimer's Disease Neuroimaging Initiative Decoding Cognitive Health Using Machine Learning: A Comprehensive Evaluation for Diagnosis of Significant Memory Concern. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery . 2024;14(5) doi: 10.1002/widm.1546.e1546 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data sharing is not applicable to this article as no new data were created or analyzed in this study.
