Skip to main content
PLOS One logoLink to PLOS One
. 2024 May 31;19(5):e0303670. doi: 10.1371/journal.pone.0303670

DAU-Net: Dual attention-aided U-Net for segmenting tumor in breast ultrasound images

Payel Pramanik 1,#, Ayush Roy 2,#, Erik Cuevas 3, Marco Perez-Cisneros 4,*, Ram Sarkar 1
Editor: Chenchu Xu5
PMCID: PMC11142567  PMID: 38820462

Abstract

Breast cancer remains a critical global concern, underscoring the urgent need for early detection and accurate diagnosis to improve survival rates among women. Recent developments in deep learning have shown promising potential for computer-aided detection (CAD) systems to address this challenge. In this study, a novel segmentation method based on deep learning is designed to detect tumors in breast ultrasound images. Our proposed approach combines two powerful attention mechanisms: the novel Positional Convolutional Block Attention Module (PCBAM) and Shifted Window Attention (SWA), integrated into a Residual U-Net model. The PCBAM enhances the Convolutional Block Attention Module (CBAM) by incorporating the Positional Attention Module (PAM), thereby improving the contextual information captured by CBAM and enhancing the model’s ability to capture spatial relationships within local features. Additionally, we employ SWA within the bottleneck layer of the Residual U-Net to further enhance the model’s performance. To evaluate our approach, we perform experiments using two widely used datasets of breast ultrasound images and the obtained results demonstrate its capability in accurately detecting tumors. Our approach achieves state-of-the-art performance with dice score of 74.23% and 78.58% on BUSI and UDIAT datasets, respectively in segmenting the breast tumor region, showcasing its potential to help with precise tumor detection. By leveraging the power of deep learning and integrating innovative attention mechanisms, our study contributes to the ongoing efforts to improve breast cancer detection and ultimately enhance women’s survival rates. The source code of our work can be found here: https://github.com/AyushRoy2001/DAUNet.

Introduction

As per the World Health Organization (WHO), breast cancer stands as the most frequently diagnosed cancer and is the primary cause of cancer-related fatalities in women globally. In the year 2020, approximately 2.3 million new instances of breast cancer were detected, constituting around 11.7% of all cancer cases worldwide. It also caused approximately 685,000 deaths, representing 6.9% of all cancer-related deaths globally [1]. Therefore, early and accurate detection of breast cancer is essential for improving treatment outcomes and patient survival rates. Among various medical imaging techniques, ultrasound imaging has played an important role in the detection and diagnosis owing to its non-invasive nature, and capability to capture high-resolution images of breast tissue [2]. However, the precise segmentation of breast lesions in ultrasound images remains a challenging task because of the presence of speckle noise, poor image quality, and inherent variations in breast tissue [3]. Manual segmentation of ultrasound images is a time-consuming task. As a result, the implementation of automatic segmentation techniques becomes important to enhance efficiency and minimize unnecessary delays [4]. Sample breast ultrasound images are shown in Fig 1.

Fig 1. Sample breast ultrasound images of benign, malignant, and normal types.

Fig 1

In recent times due to the advancements in deep learning techniques, medical image analysis and CAD systems have undergone a revolutionary transformation. There are many CAD systems that aim to automate the process of breast lesion segmentation in ultrasound images, assisting radiologists in accurate diagnosis. For instance, the popularity of Convolutional Neural Networks (CNNs) has increased for automatic detection and segmentation of breast cancer in ultrasound images. These models can learn feature representations from raw pixel intensities, enabling them to capture complex patterns and discriminative features [5]. Further, to extract the contextual information, authors of [6] introduced the U-Net model. The basic U-Net model utilizes an encoder-decoder structure, with the encoder capturing context from the input image, and the decoder generating a segmentation map through upsampling. This architecture includes skip connections between encoder and decoder layers to preserve spatial information at multiple scales. Since its introduction, U-Net has served as a foundation for several variants and extensions. Variants of U-Net include the Residual U-Net [7], which incorporates residual blocks to facilitate gradient flow and address vanishing gradient issues, the Attention U-Net [8], which integrates attention mechanisms to selectively focus on informative regions, the U-Net++ [9], which introduces nested skip connections to capture more comprehensive contextual information, the Hybrid U-Net, which combines U-Net with other architectures like VGG or ResNet for improved performance and many more [10]. These variants of U-Net have demonstrated advancements in image segmentation tasks. In a variety of computer vision tasks, attention mechanisms have produced promising results. The Attention U-Net, for instance, introduces attention gates to emphasize relevant features and suppress irrelevant ones. Other attention mechanisms, such as Squeeze-and-Excitation (SE), Channel-Attention Mechanism (CAM) and Spatial-Attention Mechanism (SAM) have also been applied to improve feature representations in CNNs [11, 12]. While U-Net and its variants excel in capturing local context, integrating global contextual information has been explored as a means to improve segmentation accuracy. Studies such as [1316] have introduced different models to capture global contextual information effectively.

Contributions

In light of the existing literature, we propose a novel segmentation method for detecting tumors in breast ultrasound images. Our approach applies two attention mechanisms, namely the Positional Convolutional Block Attention Module (PCBAM) and the Shifted Window Attention (SWA), which effectively capture context-aware features, spatial relationships, and global contextual information. The entire architecture of the proposed segmentation model is shown in Fig 2.

Fig 2. Block diagram of the proposed DAU-Net model used for segmentation of tumor in breast ultrasound images.

Fig 2

An input image with dimensions 128 × 128 × 1 undergoes feature extraction through the encoder, and the decoder then performs upsampling on the encoded features to predict a binary mask of size 128 × 128 × 1. The in-between connections of the encoder and the decoder are accompanied by the addition of PCBAM and SWA attention mechanisms to enhance the performance.

The highlights of this work are as follows:

  • Our proposed model uses two attention mechanisms in the U-Net model, called PCBAM and SWA.

  • PCBAM, the Position attention aided CBAM, combines the CBAM attention mechanism, which captures context-aware features through channel attention and spatial attention, with positional attention. This integration enhances the contextual information and spatial relationships within local features, leading to more robust and accurate representations.

  • SWA is used in the bottleneck layer of the Residual U-Net model to capture global contextual information.

  • The combination of PCBAM and SWA significantly improves the performance of the model on both the BUSI and UDIAT datasets.

The paper is organized as follows. First, we review the segmentation methods used in breast cancer ultrasound imaging by various researchers and identify the existing gaps in the literature. Next, we discuss the preliminary details and the proposed model for tumor region segmentation in breast ultrasound images. The evaluation of the model using various metrics and the analysis of the results are discussed then. Finally, we conclude our work with some potential future research directions.

Related work

Breast cancer is a pressing global health issue, and hence researchers have been exploring various methods to improve early detection, and precise diagnosis to enhance survival rates for affected women [1720]. Among these efforts, deep learning-based CAD systems have shown great promise to address this challenge. This section presents a review of relevant literature that centers around deep learning-based segmentation methods for detecting tumor regions in breast ultrasound images. Deep learning methods, especially CNNs, have shown impressive achievements in diverse medical imaging tasks, such as image segmentation, classification, and detection [2, 2123]. Researchers have applied CNNs to analyze breast ultrasound images to detect abnormalities and tumors [2426]. Studies such as [2730] explored different architectures and attention mechanisms to improve the performance of tumor segmentation in breast ultrasound images. In the study by Vakanski et al. [27], the authors combined visual saliency into a U-Net model. By incorporating visual saliency maps that capture regions attracting radiologists’ attention and combining topological and anatomical prior knowledge, the model learned feature representations prioritizing essential spatial regions. However, a limitation of this approach lies in its reliance on the quality of saliency maps, as using low-quality maps may not enhance results and could potentially lead to degraded performance. In another study by Lee et al. [30], the authors proposed a semantic segmentation network to enhance the accurate segmentation of regions of breast tumors in ultrasound images. They achieved this improvement by integrating a channel attention module with multi-scale grid average pooling (MSGRAP). This attention module enables the utilization of both global and local spatial information from input images, thereby enhancing the network’s effectiveness in performing semantic segmentation. Chen et al. introduced AAU-Net [31], which is a hybrid adaptive module, combining convolutional layers with varying kernel sizes, channel self-attention, and spatial self-attention blocks to replace the traditional convolution operation. In contrast, in another study [32], the authors introduce a cascaded CNN, which integrates U-Net, Bidirectional Attention Guidance Network (BAGNet), and Refinement Residual Network (RFNet). CBAM introduced by Woo et al. [33] demonstrated significant potential in improving the capability of CNNs to focus on relevant image regions. By integrating both channel and spatial attention mechanisms, CBAM enhances the representational power of CNNs and boosts performance in various computer vision tasks [34]. Researchers have utilized CBAM attention in tumor detection from breast cancer imaging [3537]. In [37], the introduced method employed a deep ResNet architecture with a CBAM attention module to extract more comprehensive and in-depth features from pathological images. In [35], the authors introduced a semi-supervised learning model named BUS-GAN, comprising two networks: BUS-S for segmentation and BUS-E for evaluation. The BUS-S network extracts features of multi-scale to handle variations in breast lesions, enhancing segmentation robustness. To enhance discriminative ability, the BUS-E network incorporates a dual-attentive-fusion block with spatial attention paths, distilling geometrical and intensity-level information from both the segmentation map and the original image. Through adversarial training, the BUS-GAN model achieves higher segmentation quality as the BUS-E network guides BUS-S in generating more precise segmentation maps that align closely with the ground truth distributions. Another study by Fan et al. [36] showed an approach called the Multi-Task Learning (MTL) approach to address joint breast tumor lesion localization and classification. The model comprises a classifier, an auxiliary lesion-aware network, and a shared feature extractor. Multiple attention modules are incorporated in the auxiliary network to optimize the multi-scale intermediate feature maps, and enhance representativeness through channel and spatial attention focused on lesion regions. Positional attention mechanisms have gained attention in the medical imaging domain due to their ability to capture spatial relationships and contextual information within local features. The authors of [38] utilized positional attention in a multi-scale framework to identify anatomical structures in medical images. SWA, a recent innovation proposed by [39], enhances the efficiency and adaptability of the attention mechanism. By applying the attention mechanism in a sliding window fashion, SWA effectively captures relevant information across multiple scales, making it suitable for object detection and segmentation tasks. On the other hand [40], investigated the effectiveness of an ensemble of Swin transformers for two-class (benign vs. malignant) and eight-class (four benign and four malignant sub-types) classification in medical imaging, using the BreaKHis histopathology dataset. Swin transformer is a variant of vision transformer that utilizes non-overlapping SWA. Another approach, presented by [41], introduced the BTS-ST network, which combines Swin-transformer with CNN-based U-Net for breast tumor segmentation and classification. The BTS-ST network incorporates SWA to enhance feature representation capability for irregularly shaped tumors. The residual U-Net architecture, introduced in [7], is a variant of the traditional U-Net model. Incorporating residual connections allows for better information flow during training and helps mitigate the vanishing gradient problem, leading to improved convergence and performance. Several studies have explored breast tumor segmentation using residual U-Net-based deep-learning techniques. For instance, the authors of [42] presented the RCA-IUnet, a deep-learning model designed for breast tumor segmentation in ultrasound imaging. The model integrates U-Net architecture with residual inception depth-wise separable convolution, hybrid pooling, and cross-spatial attention filters in long skip connections, effectively extracting tumor-related features. The authors conducted an ablation study, highlighting the pivotal role of residual inception convolution and cross-spatial attention components in the proposed model. However, a limitation of the model is the absence of a channel attention filter, which may restrict its capacity to emphasize the most critical feature layers. In another work reported in [43], an improved U-net MALF model was proposed for breast tumor segmentation in ultrasound images. This model enhances the attention U-net network framework by incorporating residual convolution and extended residual convolution modules in the encoding path. In their work [44], utilized a residual U-Net for breast tumor segmentation, incorporating a fusion attention mechanism that combines both spatial and channel attention. In another work [45], the authors presented the RDAU-NET (Residual-Dilated-Attention-Gate-UNet) model for tumor segmentation in breast ultrasound images. The model extends the conventional U-Net architecture and includes three modules: Residual unit, Dilation unit, and Attention Gate. These modules are introduced to improve the model’s performance and capabilities for accurate segmentation of breast tumors in ultrasound images. Several comparative studies have evaluated different deep learning-based models along with attention mechanisms for breast tumor segmentation in ultrasound images. The authors of [27, 46, 47] compared the performance of various CNN architectures with attention mechanisms, showing the potential of attention-based methods in improving segmentation accuracy. In summary, while deep learning-based methods have shown promise in breast tumor segmentation, there is still a need to explore more advanced attention mechanisms further to improve the accuracy and robustness of the models.

Methodology

Our research showcases a novel methodology that integrates PCBAM and SWA with the Residual U-Net design. This design consists of two crucial elements, the encoder and decoder, that collaborate to extract significant attributes from input images and produce accurate segmented results.

Encoder

To extract hierarchical features from the input data, the encoder employs convolutional layers with 3 × 3 filters and a stride of 1. Batch normalization and ReLU activation are applied after each convolutional layer to maintain feature stability and increase information flow. To ensure smooth gradient flow during training and to retain essential information, residual connections are utilized. To downsample, 2 × 2 stride convolutional layers are employed. Additionally, the PCBAM mechanism improves the encoder features before connecting them with the decoder features via residual connections. More information on this mechanism can be found in the subsequent Section.

Decoder

In the process of upsampling and reconstructing segmented output, the decoder plays a crucial role. By combining upsampled feature maps with those from the attention-aided encoder features, it gains access to both low-level and high-level features. This happens through strategic fusion, which involves refining the features with convolutional layers, followed by batch normalization and ReLU activation. As a result, a higher-dimensional representation of the spatial relationships is obtained. To further enhance the spatial dimensions, the decoder uses residual blocks, which contribute to its exceptional performance. Moreover, the SWA layer is incorporated in the decoder, capturing global dependencies and improving spatial coherence in the segmentation results.

Positional convolutional block attention module

CBAM attention mechanism [33] is applied to the last feature map of dimension C × H × W generated from any CNN architecture. Here, C, H and W represent a feature map’s number of channels, height, and width, respectively. The CBAM attention mechanism consists of two components: the 1D Channel Attention Module (CAM) and the 2D Spatial Attention Module (SAM). The CAM assigns weights to the channels of the feature map, enhancing specific channels that contribute more to improve model performance. It is formulated as per Eq 1.

Fc=σ(mlp(gap(F))+mlp(gmp(F))) (1)

In Eq 1, σ represents the sigmoid activation function, gap is the global average pooling layer, gmp is the global max pooling layer, and mlp denotes the multi-layer perceptron consisting of two successive fully connected i.e., dense layers (DL) with C and C/8 units, respectively and F is the feature map. Now, Fc=FcF is fed to the SAM (⊗ denotes the element-wise matrix multiplication).

The SAM operates on the feature map Fc obtained from the CAM. It applies a spatial attention mask to enhance the feature representation. The SAM is formulated according to Eq 2.

F=f7×7[DL(gap(Fc));DL(gmp(Fc))] (2)

In Eq 2, f7×7 is a convolutional layer with a kernel size of 7 × 7 and dilation of 4, DL represents the dense layers, and ‘;’ denotes the concatenation operation. The final output feature map of the CBAM attention module, denoted as FCBAM, is obtained by element-wise multiplication between F and Fc as shown in Eq 3.

FCBAM=FFc (3)

The CBAM attention mechanism effectively captures channel-wise and spatial-wise dependencies, thereby allowing the model to focus on relevant features, and improve its performance in image segmentation.

Similarly, the Position Attention Module (PAM) is designed to enrich local features by incorporating a broader context, thereby enhancing their representational capacity. To achieve this, we start with a local feature map denoted as FRH×W×C. This feature map is processed through a convolutional layer, resulting in two new feature maps, B and Z, both of size RH×W×C. Afterward, B and Z are reshaped into matrices of size RN×C, where N = H × W, representing the number of pixels in the feature map. A matrix multiplication is performed between the transpose of Z and B, followed by the application of a softmax layer, which yields the spatial attention map SRN×N. This attention map captures the spatial relationships between different pixels in the feature map. PAM allows local features to leverage a wider contextual understanding by employing the attention mechanism to emphasize relevant spatial information. This enables the local features to better represent complex patterns and structures in the input data. The formula is shown in Eq 4.

sji=exp(Bi·Zj)i=1Nexp(Bi·Zj) (4)

where sji measures the impact of the ith position on the jth position.

Next, we feed feature map F into a convolutional layer to generate a new feature map DRH×W×C, which is reshaped to RN×C. We perform a matrix multiplication between D and the transpose of S, resulting in a feature map of size RN×C. We then reshape this back to RH×W×C. Finally, we multiply it by a scale parameter α and perform an element-wise sum operation with the features F to obtain FPAMRH×W×C. The calculation is done in accordance with Eq 5.

FPAMj=αi=1N(sji·Di)+Fj (5)

where α is initialized as 0. The model learns α and gradually learns to assign more weight. The resulting feature FPAM at each position is a weighted sum of the features across all positions and original features, allowing for a global contextual view and selective aggregation of contexts based on the spatial attention map. This promotes intra-class compactness along with semantic consistency within the feature representations.

Utilizing the power of CBAM and PAM, we combine these two modules using Eq 6 to formulate the PCBAM, where the input feature to both CBAM and PAM is F. The block diagram of the PCBAM is shown in Fig 3.

FPCBAM=FPAM+FCBAM (6)

Fig 3. An illustration of the PCBAM attention block.

Fig 3

CBAM and PAM are applied to the input feature F. The addition of the outputs of CBAM and PAM is the output of the PCBAM attention mechanism, FPCBAM.

Shifted window attention

The SWA [39] is a powerful attention mechanism used to capture global dependencies and improve spatial coherence in the segmentation results of our proposed model. It enhances the model’s ability to focus on relevant regions and strengthens its contextual understanding of the input images. In image segmentation tasks, understanding the contextual relationships among different regions is crucial. However, traditional convolutional operations might not fully capture these long-range dependencies. To this end, we use the SWA module in order to address this limitation by introducing a window-based attention mechanism, which allows the model to attend to relevant information from different parts of the image.

The SWA mechanism can be mathematically defined as follows. Let F be the input feature map of size H × W × C, where H, W, and C represent the height, width, and number of channels, respectively. To compute the attention map, we first obtain position-aware query matrix q, key matrix k, and value matrix v as follows:

qi,j=Fi,j·wq (7)
ki,j=Fi,j·wk (8)
vi,j=Fi,j·wv (9)

where wq, wk, and wv are learnable weight matrices for query, key, and value projections, respectively.

Next, we perform a convolution operation f1×1 (1 × 1 is the kernel dimension) on q, k, and v to compute the attention map A as per Eq 10.

A=f1×1(q,k,v) (10)

The attention map A is then added element-wise to the original feature map F using a residual connection to obtain the final output of the SWA mechanism Xout using the following Eq 11.

Xout=F+A (11)

The SWA mechanism is integrated into the decoder part of the Residual U-Net architecture. By introducing SWA, the model can effectively capture long-range dependencies and achieve better spatial coherence in the segmentation results, leading to improved performance in segmenting breast tumor regions in ultrasound images.

Loss function

The Dice loss [48], Binary Cross Entropy (BCE) [49] loss, and Focal loss [50] are popular loss functions in image segmentation tasks. These loss functions help guide the training process of segmentation models by quantifying the similarity between the ground truth and the predicted masks.

The Dice loss is derived from the Dice Coefficient, also known as the F1 Score. It measures the overlap between the ground truth and the predicted masks, aiming to maximize their similarity. The Dice loss is computed using Eq 12.

DiceLoss=1-2×TP2×TP+FP+FN (12)

In Eq 12, TP represents the number of true positives (correctly identified foreground) pixels, FP represents the number of false positives (incorrectly identified foreground) pixels, and FN represents the number of false negatives (missed foreground) pixels.

The BCE loss is another widely used loss function for image segmentation. It measures the dissimilarity between the predicted and ground truth masks, aiming to minimize their difference. The BCE loss is computed using the following Eq 13.

BCELoss=-1Ni=1N(yilog(pi)+(1-yi)log(1-pi)) (13)

In Eq 13, N represents the total number of pixels, yi represents the ground truth label (foreground or background) for pixel i, and pi represents the predicted probability of the foreground class for pixel i.

The Focal loss is designed to address class imbalance in segmentation tasks and provide more focus on hard-to-classify pixels. It assigns higher weights to misclassified pixels and thus reduces the impact of easy-to-classify pixels during training. The Focal loss is computed using the following Eq 14.

FocalLoss=-1Ni=1N(α(1-pi)γlog(pi)) (14)

In Eq 14, α is a balancing parameter to control the contribution of each class, and γ is the focusing parameter that modulates the rate at which the loss focuses on hard-to-classify pixels.

In our model training, we have used a combination of the Dice loss, BCE loss, and Focal loss as shown in Eq 15 to guide the optimization process. By minimizing this combined loss during training, our model learns to accurately segment the desired regions of interest.

Loss=DiceLoss+BCELoss+FocalLoss (15)

Statement of ethical approval

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki Declaration and its later amendments.

Results and discussion

Dataset description

The BUSI dataset [51] includes ultrasound images from 600 female patients aged 25 to 75 years, collected at Baheya Hospital in Cairo, Egypt. The dataset contains 437 benign cases, 210 malignant cases, and 133 images of normal breast tissue, comprising a total of 730 diverse breast ultrasound images for research purposes. The images are in PNG format, and on average, it has a size of 500 × 500 pixels. In our research, we have utilized various benign and malignant cases, along with their corresponding masks, to both train and test our segmentation model. These masks are crucial in identifying areas of interest in ultrasound images and serve as ground truth annotations. Our evaluation process involves comparing the model’s predictions with the actual target regions to gauge its performance.

Evaluation metrics

The performance of our segmentation model is evaluated using commonly used metrics: Dice score, Intersection over Union (IoU) score, accuracy, recall, and precision. These metrics provide quantitative measures of the model’s ability to accurately delineate regions of interest.

Accuracy

The accuracy metric assesses the overall correctness of binary segmentation and is calculated as the ratio of correctly classified pixels to the total number of pixels. It is defined in Eq 16.

Accuracy=TP+TNTP+TN+FP+FN (16)

Precision

Precision evaluates the fraction of true positive predictions among all positive predictions and is defined in Eq 17.

Precision=TPTP+FP (17)

Recall

Recall, commonly referred to as sensitivity or true positive rate, quantifies the proportion of true positive predictions out of all the actual positive instances and is defined in Eq 18.

Recall=TPTP+FN (18)

Intersection over Union

IOU is a measure that quantifies the overlap between the ground truth mask and the predicted binary segmentation mask. It is calculated as the ratio of the intersection area between the two masks to their union area and is defined in Eq 19.

IoU=TPTP+FP+FN (19)

Dice score

The Dice score, also referred to as the F1 score, integrates both precision and recall into a single value for evaluation and is defined in Eq 20.

Dicescore=2×TP2×TP+FP+FN (20)

Experimental setup

We have developed our segmentation model using Python and have leveraged the TensorFlow and Keras libraries for implementation. For data manipulation and preprocessing, we have utilized numpy, OpenCV, and scikit-learn libraries, which have facilitated the efficient handling of data. To speed up training and utilize hardware acceleration, we use the high-performance NVIDIA TESLA P100 GPU.

Hyperparameter details

Our model is trained for 50 epochs, where every epoch represents one complete pass through the entire dataset. To address the issue of non-uniform sizes in the original BUSI images, we resize all images to a uniform size of 128 × 128 pixels, which are input into the model for segmentation. In the architecture’s convolutional layers, we utilize the ‘He Normal’ weight initialization, which has proven to be effective in deep neural network architectures. This initialization strategy contributes to better convergence and performance during training. During the training phase, we use the Adam optimizer with a learning rate of 0.0001 to optimize the model. This choice of optimizer allows us to efficiently update the model’s parameters, enhancing convergence during training. To ensure a comprehensive assessment, we have divided the data into a 70-10-20% train-test-validation split. The model is trained using 70% of the data, while the remaining subsets are reserved for testing and validation purposes. We have leveraged the training set to optimize the model’s parameters, fine-tune hyperparameters using the validation set, and gauge the model’s ability to generalize to new data by utilizing the testing set.

Ablation study

A series of experiments are conducted to refine our segmentation model and evaluate the impact of various modifications. These experiments include:

  • (i) Base Residual U-Net model, serving as the initial benchmark.

  • (ii) Residual U-Net model with PAM applied to the skip connections.

  • (iii) Residual U-Net model with CBAM applied to the skip connections.

  • (iv) Residual U-Net model with PCBAM, combining the strengths of PAM and CBAM.

  • (v) Proposed model with PCBAM and SWA, emphasizing global features.

Results in Table 1 showcase the efficacy of each modification. Each model has been trained using the linear combination of Dice, BCE, and Focal loss. The addition of PAM and CBAM improves performance, while SWA further enhances accuracy and segmentation quality.

Table 1. Performance metrics of the segmentation models.

All values are in %. Bold values indicate superior performance. The results are in xy) format, where x is the mean and y is the standard deviation of the evaluation metric for the five runs of the model.

Model Dice Accuracy Precision Recall IoU
(i) 68.27 (±0.60) 92.85 (±0.60) 68.15 (±0.41) 71.71 (±0.77) 55.82 (±0.53)
(ii) 71.72 (±0.32) 94.55 (±0.27) 78.28 (±0.16) 65.45 (±0.39) 61.08 (±0.49)
(iii) 72.08 (±0.50) 94.38 (±0.25) 74.42 (±0.39) 70.86 (±0.59) 62.35 (±0.24)
(iv) 71.97 (±0.39) 94.26 (±0.45) 73.76 (±0.58) 69.66 (±0.76) 62.99 (±0.43)
(v) 74.23 (±0.67) 95.88 (±0.42) 73.81 (±0.43) 74.59 (±0.65) 65.32 (±0.56)

Fig 4 visually illustrates the performance enhancement achieved with attention mechanisms. The combination of PCBAM and SWA results in improved performance for both the small and large region of interest, refining feature representations and capturing both global and local spatial dependencies for accurate segmentation.

Fig 4. Results of the ablation study indicate the improvement in model performance with each experimental modification.

Fig 4

GT and PM are Ground Truth and Predicted Mask, respectively. Fc is the heatmap of the bottleneck layer and it demonstrates the improvement of the model’s performance in focusing on the region of interest after the addition of the SWA in the bottleneck layer. Fa and Fb are heatmaps of the features flowing from the first and second encoder layers to the first and second decoder layers via skip connections. It can be seen that Fa and Fb get more enriched with the use of attentions such as CBAM, PAM, and PCBAM.

Through these experiments and analyses, we are able to improve our segmentation model iteratively, identifying the most effective modifications and attention mechanisms. These advancements make a significant contribution to enhancing the accuracy and robustness of our model, positioning it as an advanced solution for segmenting breast tumors in ultrasound images. Additionally, we experiment with a five-fold cross validation [52] approach for assessing the model’s generalizability, and tabulate the results under Table 2.

Table 2. Results of the proposed DAU-Net model with 5-fold cross-validation on the BUSI dataset.

5-Fold CV IoU(%) DSC(%)
Fold 1 64.85 73.58
Fold 2 65.43 74.14
Fold 3 65.17 74.29
Fold 4 64.79 73.21
Fold 5 65.91 74.61
Mean 65.23 73.97
Std. Dev. 0.411 0.504

Statistical analysis

We have conducted a statistical test to assess the robustness of the proposed segmentation model compared to the other models considered in the ablation study. We hypothesize that “The proposed DAU-Net model yields similar results in comparison to the other models considered in the ablation study.” To perform this test, we have considered the Mann-Whitney U test [53], a popular non-parametric statistical test. We have compared the Dice and IoU scores from five different runs for each model, as described in the last section (i, ii, iii, and iv) of the base models, with the proposed model (v) to perform this analysis. The results are presented in Table 3. We can safely reject the null hypothesis for each case based on the results provided in Table 3 because the p-value is less than 0.05 (5%) in each case. Furthermore, we have noted that the magnitudes of the results are identical. However, as the Mann-Whitney U test is rank-based and is not dependent on the magnitude of the results, this characteristic does not affect the validity of the statistical test. In conclusion, the statistical analysis using the Mann-Whitney U test provides strong evidence that the proposed DAU-Net model yields statistically significant results compared to the other models considered in the ablation study. This suggests that the use of the dual attention methodology in the present work contributes to the model’s effectiveness and reliability.

Table 3. Results of the Mann-Whitney U test of the proposed DAU-Net model used for segmenting tumor regions in breast images of the BUSI dataset.

Model p-value (Dice) p-value (IoU)
(i) 0.007 0.007
(ii) 0.007 0.007
(iii) 0.007 0.007
(iv) 0.011 0.011

Additional experimentation

We have performed a series of experiments utilizing different loss combinations and have consolidated the results in Table 4. This table highlights the ablation study we have conducted on the loss functions employed to train our model. After analyzing Table 4, we have determined the most efficient loss function during model training and selected the optimal model configuration. Our findings reveal that the combination of Dice, BCE, and focal loss yields the highest performance.

Table 4. Performance metrics of the proposed model with different loss functions.

Loss Accuracy Dice Precision Recall IoU
BCE loss 94.50 56.73 53.64 65.18 48.92
Dice loss 94.69 58.94 53.01 67.79 50.31
Focal loss 94.13 53.16 76.40 49.08 43.40
BCE loss + Dice loss 95.04 73.32 77.62 72.69 63.05
BCE loss + Focal loss 94.97 70.61 77.48 65.63 58.27
Dice loss + Focal loss 94.51 69.35 72.53 70.13 54.81
Dice loss + Focal loss + BCE loss 95.88 74.23 73.81 74.59 65.32

State-of-the-art comparison

We have performed a comparison between our proposed model and several state-of-the-art (SOTA) models along with standard segmentation models. The comparative results, comprehensively evaluating various evaluation metrics, are presented in Tables 5 and 6. The models compared in Table 5 are well-known in the image segmentation field, such as FCN [54], U-Net [6] SegNet [55] and ENC-Net [56]. Our proposed method demonstrates superior performance across the standard models (see Table 5) in terms of Dice score, Precision, and IoU, indicating better overall segmentation accuracy. Furthermore, our proposed model outperforms other advanced models (see Table 6, including ResUNet++ [57], SCAN [58], STAN [59], ColonSegnet [60] and AE-Unet [61], in terms of Dice score and IoU, highlighting its ability to accurately capture the overlap between predicted and ground truth segmentation masks. However, it is important to note that the precision and recall values for our proposed method are slightly lower than some of these models, suggesting a potential trade-off between precision and recall.

Table 5. Performance comparison with standard segmentation models.

All values are in %. Bold values indicate superior performance.

Model Dice Precision Recall IoU
FCN [54] 71.23 69.07 77.02 56.27
UNet [6] 71.32 66.96 78.46 56.13
SegNet [55] 72.25 68.77 80.06 60.01
ENC-Net [56] 72.66 68.59 79.90 57.70
Proposed 74.23 73.81 74.59 65.32

Table 6. Performance comparison with SOTA models.

All values are in %. Bold values indicate superior performance.

Model Dice Precision Recall IoU
DA-Net [62] 67.83 - 80.38 -
ResUNet++ [57] 73.85 80.10 71.43 60.02
SK-UNET [63] 70.90 - 80.80 -
SCAN [58] 72.00 73.00 - -
STAN [59] 72.00 76.00 - -
ColonSegnet [60] 73.53 76.81 76.43 62.71
MCF-Net [64] 71.06 - 72.23 -
UNext [65] 65.94 - - 55.22
AE-Unet [61] 73.47 74.44 79.00 64.57
RRC-Net [66] 72.53 71.73 77.72 63.60
MBSNet [67] 72.81 - - 63.21
U-Net-densenet121 [68] 73.70 - 72.55 62.46
Proposed 74.23 73.81 74.59 65.32

To provide specific performance details, our proposed model achieves a Dice score of 74.23, indicating a higher level of similarity between the ground truth and the predicted segmentation masks. Additionally, precision with a value of 73.81 indicates a significant proportion of predicted foreground pixels are indeed correctly identified. The Recall value of 74.59 showcases the model’s ability to accurately identify a substantial number of actual foreground pixels. Moreover, the IoU metric, with a value of 65.32, indicates the model’s strong capability in accurately delineating regions of interest. Significantly, the proposed method achieves the highest Dice score and IoU out of all the models listed in Table 6, suggesting that it excels in terms of segmentation accuracy and overlap with the ground truth.

Overall, the evaluation results showcase the superior performance of our proposed model compared to the state-of-the-art models. This confirms the effectiveness and robustness of our model in achieving accurate and precise segmentation results, positioning it as a promising solution for various segmentation tasks. However, the slightly lower Precision and Recall value compared to some models may indicate a potential area for improvement. Fig 5 showcases the segmentation results of our proposed model, demonstrating its ability to segment breast tumor regions in ultrasound images accurately. The heatmaps showcase the spatial regions where the SWA and PCBAM layers focus. Furthermore, the heatmap visualization of the proposed model as shown in Fig 4 illustrates the spatial regions where it places more emphasis, showing a close resemblance to the ground truth regions for the BUSI dataset. This indicates that the model focuses its attention on relevant areas, contributing to its accurate segmentation performance.

Fig 5. Results of the proposed segmentation model on images of the BUSI dataset and the heatmaps of SWA and PCBAM layers.

Fig 5

PCBAM1 corresponds to the PCBAM layer just above the SWA layer, PCBAM2 corresponds to the PCBAM layer just above PCBAM1 layer, and PCBAM3 corresponds to the PCBAM layer just above PCBAM2 layer.

Through these comprehensive evaluations and visualizations, our proposed model showcases its potential to significantly improve breast cancer detection and diagnosis, bringing us closer to the goal of early detection and enhanced patient care.

Error analysis

Our proposed model has demonstrated excellent performance across various image segmentation tasks, outperforming SOTA models, as depicted in Tables 5 and 6. It is essential to highlight that the precision and recall are relatively lower, indicating instances where non-tumorous regions are misclassified as tumorous and vice-versa. It is important to acknowledge the complexity of the dataset used for evaluation, which presents challenges in achieving perfect segmentation results. Fig 6 illustrates specific cases where our model encounters difficulties, resulting in deviations from the ground truth segmentation. These challenges may arise from dataset complexity, variations in image quality, or the presence of ambiguous features that are hard to accurately delineate.

Fig 6. Illustration of some of the failed cases of our model.

Fig 6

The encircled regions are the misclassified segmented masks. GT and PM represent the Ground Truth and Predicted Mask, respectively.

Despite these challenges, our proposed model demonstrates significant potential and promises a valuable contribution to the field of breast cancer detection. By continuing to address the limitations and exploring further research directions, we aim to enhance the model’s segmentation performance and make strides toward more accurate and reliable breast cancer diagnosis.

Experimentation on the UDIAT dataset

To evaluate the effectiveness of our proposed DAU-Net method, we have conducted assessments on the UDIAT dataset, also known as Dataset B, a well-known collection of breast ultrasound images generously provided by the UDIAT Diagnostic Centre in Sabadell, Spain [69]. This dataset comprises a total of 163 images, consisting of 109 benign and 54 malignant ultrasound images, each accompanied by its respective ground truth mask. The average resolution for both the ultrasound images and the corresponding ground truth masks is 760 × 570 pixels. For the evaluation on the UDIAT dataset, we have maintained consistency by using the same set of hyperparameters that are employed in the evaluation of the BUSI dataset. The segmentation results of the proposed method on the UDIAT dataset is shown in Fig 7. Table 7 presents a comprehensive overview of the quantitative performance achieved by our proposed model when compared to previous notable research efforts conducted on this dataset.

Fig 7. Predicted mask and heatmap visualization of the proposed model on the UDIAT dataset.

Fig 7

GT and PM represent the Ground Truth and Predicted Mask, respectively. F a, F b, and F c are the heatmaps of the features flowing from the first and second encoder layers to the first and second decoder layers via skip connections and the bottleneck layer, respectively.

Table 7. Performance comparison of the proposed model with past methods on UDIAT dataset.

All values are in %. Bold values indicate superior performance.

Model Dice Precision IoU
SegNet [55] 70.80 85.00 60.00
CE-Net [70] 72.00 74.00 61.00
MultiResUNet [71] 75.00 79.00 66.00
SCAN [58] 74.00 75.00 65.00
U-Net [6] 75.00 78.00 65.00
STAN [59] 78.20 80.00 -
Proposed 78.58 85.85 64.71

Conclusion and future scope

In conclusion, breast cancer continues to be a pressing issue worldwide, underscoring the significance of timely identification and precise diagnosis to enhance outcomes. With the latest strides in deep learning, CAD solutions have emerged as promising tools in this domain. The current study introduces a novel segmentation technique called the PCBAM attention-based Residual U-Net model for detecting breast tumors in ultrasound images. Our approach has delivered satisfactory outcomes, revealing its potential to improve breast cancer detection and diagnosis. Nonetheless, it is crucial to recognize the constraints of our proposed method.

During the error analysis section, several potential areas for future research were identified. One particularly promising direction involves investigating multi-modal architectures that integrate multiple forms of data sources within the model. Such an approach has the potential to improve performance and deepen our comprehension of complex breast cancer detection challenges by leveraging the complementary insights provided by various modalities. Another avenue for future research is to enhance datasets through techniques such as data synthesis or generation, which can increase their diversity and size, thereby enhancing generalization and robustness.

Additionally, evaluating the robustness of the model could provide valuable insights. Performing segmentation on other types of medical images, beyond breast ultrasound images, would be beneficial in assessing the model’s versatility and applicability in diverse medical imaging tasks. In conclusion, our proposed PCBAM attention-based Residual U-Net model shows promise in breast tumor detection in ultrasound images. However, continued research in multi-modal architectures, dataset augmentation, and evaluation of other medical images will contribute to the advancement of accurate and reliable breast cancer diagnosis.

Acknowledgments

We are thankful to the Center for Microprocessor Applications for Training Education and Research (CMATER) research laboratory of the Computer Science and Engineering Department, Jadavpur University, Kolkata, India, for providing infrastructural support to this research project.

Data Availability

The publicly available dataset is analyzed in this study. The data can be found here: [URL: 1. BUSI: https://www.kaggle.com/datasets/aryashah2k/breast-ultrasound-images-dataset (accessed on 20th February, 2023) 2. UDIAT: http://www2.docm.mmu.ac.uk/STAFF/M.Yap/dataset.php.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1. Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: a cancer journal for clinicians. 2018;68(6):394–424. [DOI] [PubMed] [Google Scholar]
  • 2.Pramanik P, Mukhopadhyay S, Kaplun D, Sarkar R. A deep feature selection method for tumor classification in breast ultrasound images. In: International conference on mathematics and its applications in new computer systems. Springer; 2021. p. 241–252.
  • 3. Muhammad M, Zeebaree D, Brifcani AMA, Saeed J, Zebari DA. Region of interest segmentation based on clustering techniques for breast cancer ultrasound images: A review. Journal of Applied Science and Technology Trends. 2020;1(3):78–91. doi: 10.38094/jastt20201328 [DOI] [Google Scholar]
  • 4. Huang Q, Luo Y, Zhang Q. Breast ultrasound image segmentation: a survey. International journal of computer assisted radiology and surgery. 2017;12:493–507. doi: 10.1007/s11548-016-1513-1 [DOI] [PubMed] [Google Scholar]
  • 5. Gómez-Flores W, de Albuquerque Pereira WC. A comparative study of pre-trained convolutional neural networks for semantic segmentation of breast tumors in ultrasound. Computers in Biology and Medicine. 2020;126:104036. doi: 10.1016/j.compbiomed.2020.104036 [DOI] [PubMed] [Google Scholar]
  • 6.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer; 2015. p. 234–241.
  • 7. Zhang Z, Liu Q, Wang Y. Road extraction by deep residual u-net. IEEE Geoscience and Remote Sensing Letters. 2018;15(5):749–753. doi: 10.1109/LGRS.2018.2802944 [DOI] [Google Scholar]
  • 8.Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, et al. Attention u-net: Learning where to look for the pancreas. arXiv preprint arXiv:180403999. 2018;.
  • 9.Zhou Z, Rahman Siddiquee MM, Tajbakhsh N, Liang J. Unet++: A nested u-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4. Springer; 2018. p. 3–11. [DOI] [PMC free article] [PubMed]
  • 10. Siddique N, Paheding S, Elkin CP, Devabhaktuni V. U-net and its variants for medical image segmentation: A review of theory and applications. Ieee Access. 2021;9:82031–82057. doi: 10.1109/ACCESS.2021.3086020 [DOI] [Google Scholar]
  • 11. Hu J , Shen L , Sun G . Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2018. p. 7132–7141.
  • 12.Fang W, Han Xh. Spatial and channel attention modulated network for medical image segmentation. In: Proceedings of the Asian Conference on Computer Vision; 2020.
  • 13. Ni J, Wu J, Tong J, Chen Z, Zhao J. GC-Net: Global context network for medical image segmentation. Computer methods and programs in biomedicine. 2020;190:105121. doi: 10.1016/j.cmpb.2019.105121 [DOI] [PubMed] [Google Scholar]
  • 14. Zhang J, Qin Q, Ye Q, Ruan T. ST-unet: Swin transformer boosted U-net with cross-layer feature enhancement for medical image segmentation. Computers in Biology and Medicine. 2023;153:106516. doi: 10.1016/j.compbiomed.2022.106516 [DOI] [PubMed] [Google Scholar]
  • 15. Amer A, Lambrou T, Ye X. MDA-unet: a multi-scale dilated attention U-net for medical image segmentation. Applied Sciences. 2022;12(7):3676. doi: 10.3390/app12073676 [DOI] [Google Scholar]
  • 16. Feng S, Zhao H, Shi F, Cheng X, Wang M, Ma Y, et al. CPFNet: Context pyramid fusion network for medical image segmentation. IEEE transactions on medical imaging. 2020;39(10):3008–3018. doi: 10.1109/TMI.2020.2983721 [DOI] [PubMed] [Google Scholar]
  • 17. Majumdar S, Pramanik P, Sarkar R. Gamma function based ensemble of CNN models for breast cancer detection in histopathology images. Expert Systems with Applications. 2023;213:119022. doi: 10.1016/j.eswa.2022.119022 [DOI] [Google Scholar]
  • 18. Pramanik R, Pramanik P, Sarkar R. Breast cancer detection in thermograms using a hybrid of GA and GWO based deep feature selection method. Expert Systems with Applications. 2023;219:119643. doi: 10.1016/j.eswa.2023.119643 [DOI] [Google Scholar]
  • 19. Pramanik P, Mukhopadhyay S, Mirjalili S, Sarkar R. Deep feature selection using local search embedded social ski-driver optimization algorithm for breast cancer detection in mammograms. Neural Computing and Applications. 2023;35(7):5479–5499. doi: 10.1007/s00521-022-07895-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Bagchi A, Pramanik P, Sarkar R. A Multi-Stage Approach to Breast Cancer Classification Using Histopathology Images. Diagnostics. 2022;13(1):126. doi: 10.3390/diagnostics13010126 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wang Z. Deep learning in medical ultrasound image segmentation: a review. arXiv preprint arXiv:200207703. 2020;.
  • 22. Sarvamangala D, Kulkarni RV. Convolutional neural networks in medical image understanding: a survey. Evolutionary intelligence. 2022;15(1):1–22. doi: 10.1007/s12065-020-00540-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Liu S, Cai T, Tang X, Wang C. MRL-Net: Multi-scale Representation Learning Network for COVID-19 Lung CT Image Segmentation. IEEE Journal of Biomedical and Health Informatics. 2023; p. 1–14. doi: 10.1109/JBHI.2023.3285936 [DOI] [PubMed] [Google Scholar]
  • 24. Ayana G, Dese K, Choe Sw. Transfer learning in breast cancer diagnoses via ultrasound imaging. Cancers. 2021;13(4):738. doi: 10.3390/cancers13040738 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Fujioka T, Mori M, Kubota K, Oyama J, Yamaga E, Yashima Y, et al. The utility of deep learning in breast ultrasonic imaging: a review. Diagnostics. 2020;10(12):1055. doi: 10.3390/diagnostics10121055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Pramanik P, Pramanik R, Schwenker F, Sarkar R. DBU-Net: Dual branch U-Net for tumor segmentation in breast ultrasound images. Plos one. 2023;18(11):e0293615. doi: 10.1371/journal.pone.0293615 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Vakanski A, Xian M, Freer PE. Attention-enriched deep learning model for breast tumor segmentation in ultrasound images. Ultrasound in medicine & biology. 2020;46(10):2819–2833. doi: 10.1016/j.ultrasmedbio.2020.06.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Jahwar AF, Abdulazeez AM. Segmentation and classification for breast cancer ultrasound images using deep learning techniques: a review. In: 2022 IEEE 18th International Colloquium on Signal Processing & Applications (CSPA). IEEE; 2022. p. 225–230.
  • 29. Chen G, Li L, Zhang J, Dai Y. Rethinking the unpretentious U-net for medical ultrasound image segmentation. Pattern Recognition. 2023; p. 109728. doi: 10.1016/j.patcog.2023.109728 [DOI] [Google Scholar]
  • 30. Lee H, Park J, Hwang JY. Channel attention module with multiscale grid average pooling for breast cancer segmentation in an ultrasound image. IEEE transactions on ultrasonics, ferroelectrics, and frequency control. 2020;67(7):1344–1353. [DOI] [PubMed] [Google Scholar]
  • 31. Chen G, Li L, Dai Y, Zhang J, Yap MH. AAU-net: an adaptive attention U-net for breast lesions segmentation in ultrasound images. IEEE Transactions on Medical Imaging. 2022;. [DOI] [PubMed] [Google Scholar]
  • 32. Chen G, Dai Y, Zhang J. C-Net: Cascaded convolutional neural network with global guidance and refinement residuals for breast ultrasound images segmentation. Computer Methods and Programs in Biomedicine. 2022;225:107086. doi: 10.1016/j.cmpb.2022.107086 [DOI] [PubMed] [Google Scholar]
  • 33.Woo S, Park J, Lee JY, Kweon IS. Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV); 2018. p. 3–19.
  • 34.Luo Y, Wang Z. An improved resnet algorithm based on cbam. In: 2021 International Conference on Computer Network, Electronic and Automation (ICCNEA). IEEE; 2021. p. 121–125.
  • 35. Han L, Huang Y, Dou H, Wang S, Ahamad S, Luo H, et al. Semi-supervised segmentation of lesion from breast ultrasound images with attentional generative adversarial network. Computer methods and programs in biomedicine. 2020;189:105275. doi: 10.1016/j.cmpb.2019.105275 [DOI] [PubMed] [Google Scholar]
  • 36. Fan Z, Gong P, Tang S, Lee CU, Zhang X, Song P, et al. Joint localization and classification of breast tumors on ultrasound images using a novel auxiliary attention-based framework. arXiv preprint arXiv:221005762. 2022;. [DOI] [PMC free article] [PubMed]
  • 37. Zhang X, Zhang Y, Qian B, Liu X, Li X, Wang X, et al. Classifying breast cancer histopathological images using a robust artificial neural network architecture. In: Bioinformatics and Biomedical Engineering: 7th International Work-Conference, IWBBIO 2019, Granada, Spain, May 8-10, 2019, Proceedings, Part I 7. Springer; 2019. p. 204–215.
  • 38. Liu C, Gu P, Xiao Z, et al. Multiscale U-Net with spatial positional attention for retinal vessel segmentation. Journal of Healthcare Engineering. 2022;2022. doi: 10.1155/2022/5188362 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF international conference on computer vision; 2021. p. 10012–10022.
  • 40. Tummala S, Kim J, Kadry S. BreaST-Net: Multi-class classification of breast cancer from histopathological images using ensemble of swin transformers. Mathematics. 2022;10(21):4109. doi: 10.3390/math10214109 [DOI] [Google Scholar]
  • 41. Iqbal A, Sharif M. BTS-ST: Swin transformer network for segmentation and classification of multimodality breast cancer images. Knowledge-Based Systems. 2023;267:110393. doi: 10.1016/j.knosys.2023.110393 [DOI] [Google Scholar]
  • 42. Punn NS, Agarwal S. RCA-IUnet: a residual cross-spatial attention-guided inception U-Net model for tumor segmentation in breast ultrasound imaging. Machine Vision and Applications. 2022;33(2):27. doi: 10.1007/s00138-022-01280-3 [DOI] [Google Scholar]
  • 43. Tong Y, Liu Y, Zhao M, Meng L, Zhang J. Improved U-net MALF model for lesion segmentation in breast ultrasound images. Biomedical Signal Processing and Control. 2021;68:102721. doi: 10.1016/j.bspc.2021.102721 [DOI] [Google Scholar]
  • 44. Zhao T, Dai H. Breast tumor ultrasound image segmentation method based on improved residual u-net network. Computational Intelligence and Neuroscience. 2022;2022. doi: 10.1155/2022/3905998 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Zhuang Z, Li N, Joseph Raj AN, Mahesh VG, Qiu S. An RDAU-NET model for lesion segmentation in breast ultrasound images. PloS one. 2019;14(8):e0221535. doi: 10.1371/journal.pone.0221535 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Luo Y, Huang Q, Li X. Segmentation information with attention integration for classification of breast tumor in ultrasound image. Pattern Recognition. 2022;124:108427. doi: 10.1016/j.patcog.2021.108427 [DOI] [Google Scholar]
  • 47. He Q, Yang Q, Xie M. HCTNet: A hybrid CNN-transformer network for breast ultrasound image segmentation. Computers in Biology and Medicine. 2023;155:106629. doi: 10.1016/j.compbiomed.2023.106629 [DOI] [PubMed] [Google Scholar]
  • 48. Soomro TA, Afifi AJ, Gao J, Hellwich O, Paul M, Zheng L. Strided U-Net model: Retinal vessels segmentation using dice loss. In: 2018 Digital Image Computing: Techniques and Applications (DICTA). IEEE; 2018. p. 1–8. [Google Scholar]
  • 49.Jadon S. A survey of loss functions for semantic segmentation. In: 2020 IEEE conference on computational intelligence in bioinformatics and computational biology (CIBCB). IEEE; 2020. p. 1–7.
  • 50.Lin TY, Goyal P, Girshick R, He K, Dollár P. Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision; 2017. p. 2980–2988.
  • 51. Al-Dhabyani W, Gomaa M, Khaled H, Fahmy A. Dataset of breast ultrasound images. Data in brief. 2020;28:104863. doi: 10.1016/j.dib.2019.104863 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Refaeilzadeh P, Tang L, Liu H. Cross-validation. Encyclopedia of database systems. 2009; p. 532–538. doi: 10.1007/978-0-387-39940-9_565 [DOI] [Google Scholar]
  • 53. Mann HB, Whitney DR. On a test of whether one of two random variables is stochastically larger than the other. The annals of mathematical statistics. 1947; p. 50–60. doi: 10.1214/aoms/1177730491 [DOI] [Google Scholar]
  • 54.Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2015. p. 3431–3440. [DOI] [PubMed]
  • 55. Badrinarayanan V, Kendall A, Cipolla R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine intelligence. 2017;39(12):2481–2495. doi: 10.1109/TPAMI.2016.2644615 [DOI] [PubMed] [Google Scholar]
  • 56.Zhang H, Dana K, Shi J, Zhang Z, Wang X, Tyagi A, et al. Context encoding for semantic segmentation. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition; 2018. p. 7151–7160.
  • 57.Jha D, Bhattacharya S, Chandra S, Kalra S, Srivastava R. ResUNet++: An Advanced Architecture for Medical Image Segmentation. In: 2019 IEEE International Symposium on Multimedia (ISM). IEEE; 2019.
  • 58.Zhang B, Lu L, Yao J, Wang X, Summers RM. Attention-based CNN for KL grade classification: Data from the osteoarthritis initiative. In: 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI). IEEE; 2020. p. 1006–1009.
  • 59.Shareef B, Xian M, Vakanski A. Stan: Small tumor-aware network for breast ultrasound image segmentation. In: 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI). IEEE; 2020. p. 1–5. [DOI] [PMC free article] [PubMed]
  • 60. Jha D, Jha A, Thangali A, Jha H, Saini D, Jha P, et al. Real-time polyp detection, localization and segmentation in colonoscopy using deep learning. IEEE Access. 2021;9:40496–40510. doi: 10.1109/ACCESS.2021.3063716 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Yan Y, Liu Y, Wu Y, Zhang H, Zhang Y, Meng L. Accurate segmentation of breast tumors using AE U-net with HDC model in ultrasound images. Biomedical Signal Processing and Control. 2022;72:103299. doi: 10.1016/j.bspc.2021.103299 [DOI] [Google Scholar]
  • 62.Sun K, Xiao B, Liu D, Wang J. Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2019. p. 5693–5703.
  • 63. Byra M, Kot M, Paja W. Breast mass segmentation in ultrasound with selective kernel U-Net convolutional neural network. Biomedical Signal Processing and Control. 2020;61:102027. doi: 10.1016/j.bspc.2020.102027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Liu L, Liu J, Zheng J, Chen S, Li H, Wang X, et al. A novel MCF-Net: Multi-level context fusion network for 2D medical image segmentation. Computer Methods and Programs in Biomedicine. 2022;226:107160. doi: 10.1016/j.cmpb.2022.107160 [DOI] [PubMed] [Google Scholar]
  • 65.Valanarasu JMJ, Patel VM. Unext: Mlp-based rapid medical image segmentation network. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2022. p. 23–33.
  • 66. Chen G, Dai Y, Zhang J. RRCNet: Refinement residual convolutional network for breast ultrasound images segmentation. Engineering Applications of Artificial Intelligence. 2023;117:105601. doi: 10.1016/j.engappai.2022.105601 [DOI] [Google Scholar]
  • 67. Jin S, Yu S, Peng J, Wang H, Zhao Y. A novel medical image segmentation approach by using multi-branch segmentation network based on local and global information synchronous learning. Scientific Reports. 2023;13(1):6762. doi: 10.1038/s41598-023-33357-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68. Bal-Ghaoui M, Alaoui MHEY, Jilbab A, Bourouhou A. U-Net transfer learning backbones for lesions segmentation in breast ultrasound images. International Journal of Electrical and Computer Engineering (IJECE). 2023;13(5):5747–5754. doi: 10.11591/ijece.v13i5.pp5747-5754 [DOI] [Google Scholar]
  • 69. Yap MH, Pons G, Marti J, Ganau S, Sentis M, Zwiggelaar R, et al. Automated breast ultrasound lesions detection using convolutional neural networks. IEEE journal of biomedical and health informatics. 2017;22(4):1218–1226. doi: 10.1109/JBHI.2017.2731873 [DOI] [PubMed] [Google Scholar]
  • 70. Gu Z, Cheng J, Fu H, Zhou K, Hao H, Zhao Y, et al. Ce-net: Context encoder network for 2d medical image segmentation. IEEE transactions on medical imaging. 2019;38(10):2281–2292. doi: 10.1109/TMI.2019.2903562 [DOI] [PubMed] [Google Scholar]
  • 71. Ibtehaz N, Rahman MS. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural networks. 2020;121:74–87. doi: 10.1016/j.neunet.2019.08.025 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Chenchu Xu

21 Dec 2023

PONE-D-23-35838DAU-Net: Dual Attention-aided U-Net for Segmenting Tumor in Breast Ultrasound ImagesPLOS ONE

Dear Dr. Perez-Cisneros,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 04 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Chenchu Xu, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that your Data Availability Statement is currently as follows: All relevant data are within the manuscript and its Supporting Information files

Please confirm at this time whether or not your submission contains all raw data required to replicate the results of your study. Authors must share the “minimal data set” for their submission. PLOS defines the minimal data set to consist of the data required to replicate all study findings reported in the article, as well as related metadata and methods (https://journals.plos.org/plosone/s/data-availability#loc-minimal-data-set-definition).

For example, authors should submit the following data:

- The values behind the means, standard deviations and other measures reported;

- The values used to build graphs;

- The points extracted from images for analysis.

Authors do not need to submit their entire data set if only a portion of the data was used in the reported study.

If your submission does not contain these data, please either upload them as Supporting Information files or deposit them to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories.

If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. If data are owned by a third party, please indicate how others may request data access.

4. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

Additional Editor Comments:

The reviewers suggest improvements in several key areas: expanding dataset validation for better generalizability, enhancing language clarity, providing more detailed descriptions of innovations like PCBAM and SWA, clarifying statistical methods and results, and updating the literature review to include the latest research. Additionally, the manuscript should address potential overfitting as indicated in Fig. 4, present comparative predictive results, apply k-fold cross-validation for fairness, and consider the implications of image resizing dimensions. These revisions are essential for enhancing the manuscript's overall quality and relevance to the field.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In my opinion, some important problems in this paper need to be thoroughly revised:

1) The proposed method is only verified on one kind of dataset, which I think is very insufficient. I hope the authors can experiment with other datasets.

2) The English expression of this paper is not very standardized, which brings some confusion to the reader's understanding. I think it could be polished by someone more professional.

3) The proposed PCBAM and SWA are the main innovations of this paper. Here, PCBAM is already a relatively conventional method, so it can be considered meaningful. However, the SWA module authors did not describe it accurately enough in to “Shifted Window Attention” section. I did not understand the specifics of the module. Moreover, the authors can refer to these papers: https://doi.org/10.1016/j.media.2023.102980, https://doi.org/10.1016/j.compmedimag.2022.102054, etc.

4) What does Fig. 4 mean? Prove that the model is overfitted. If the results on the validation set are so volatile, are the results of the test set still reliable?

5) Table 2 shows the "Mann-Whitney U test" or "p-value test". I think these are two different evaluation methods. Moreover, are all the p-values of the 4 ablation methods shown in the table and the proposed DAU-Net model 0.0286? I remain skeptical.

6) Is there a difference between “mlp“ in formula (1) and ”DL“ in formula (2)?

Reviewer #2: Breast lesion segmentation is a very interesting work. In this paper, they present a novel Dual Attention-aided U-Net (DAU-Net) for breast ultrasound segmentation. Although the study is interesting, this paper has a few concerns for publication:

1. In Abstract, it would have been interesting to quantify the results obtained at this level, to guide readers on the performance of your model.

2. The authors have omitted much of the latest research work. Especially the work of 2022 and 2023 is missing. It is recommended that the authors add the latest literature to help readers gain more comprehensive knowledge. The following work can be referred:

[1] https://doi.org/10.1109/TMI.2022.3226268

[2] https://doi.org/10.1016/j.patcog.2023.109728

[3] https://doi.org/10.1016/j.cmpb.2022.107086

3. In Fig. 6, the author needs to display the prediction results of each method.

4. To ensure the fairness of the experiment, k-fold cross validation is necessary. It can be referred to in future research.

5. Would resizing the original image to 128 be too small?

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 May 31;19(5):e0303670. doi: 10.1371/journal.pone.0303670.r002

Author response to Decision Letter 0


29 Jan 2024

We are very grateful to the Editor for considering our manuscript for review and we are very

thankful to the reviewers for providing us with valuable suggestions and giving feedback about

our manuscript. We have made the required changes to address the editor’s and reviewers’

comments. Please note that the changes are highlighted in the revised manuscript and herein as

blue for Reviewer #1 and green for Reviewer #2 for the convenience of the reviewers. Replies to

each of the reviewers’ comments are also listed below

Attachment

Submitted filename: Response sheet_PONE-D-23-35838.pdf

pone.0303670.s001.pdf (79.3KB, pdf)

Decision Letter 1

Chenchu Xu

29 Feb 2024

PONE-D-23-35838R1DAU-Net: Dual Attention-aided U-Net for Segmenting Tumor in Breast Ultrasound ImagesPLOS ONE

Dear Dr. Perez-Cisneros,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 14 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Chenchu Xu, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Is it true that the training losses and accuracy in Fig. 4 are stable, but the validation losses and accuracy are so volatile? This is a problem that I don't think has been effectively addressed.

Reviewer #2: Thanks for the effort put in as a solution to my doubt. The revised manuscript may be considered for acceptance.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 May 31;19(5):e0303670. doi: 10.1371/journal.pone.0303670.r004

Author response to Decision Letter 1


30 Mar 2024

We are very grateful to the Editor for considering our manuscript for review and we are very

thankful to the reviewers for providing us with valuable suggestions and giving feedback about

our manuscript. We have made the required changes to address the editor’s and reviewers’

comments.

______________________________________________________________________________

Attachment

Submitted filename: Response to Reviewers DAU_Net_PONE v2.pdf

pone.0303670.s002.pdf (306.4KB, pdf)

Decision Letter 2

Chenchu Xu

30 Apr 2024

DAU-Net: Dual Attention-aided U-Net for Segmenting Tumor in Breast Ultrasound Images

PONE-D-23-35838R2

Dear Dr. Perez-Cisneros,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Chenchu Xu, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The questions I raised have been revised, and the author has made a detailed reply. I think this article is acceptable.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

Acceptance letter

Chenchu Xu

21 May 2024

PONE-D-23-35838R2

PLOS ONE

Dear Dr. Perez-Cisneros,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Chenchu Xu

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response sheet_PONE-D-23-35838.pdf

    pone.0303670.s001.pdf (79.3KB, pdf)
    Attachment

    Submitted filename: Response to Reviewers DAU_Net_PONE v2.pdf

    pone.0303670.s002.pdf (306.4KB, pdf)

    Data Availability Statement

    The publicly available dataset is analyzed in this study. The data can be found here: [URL: 1. BUSI: https://www.kaggle.com/datasets/aryashah2k/breast-ultrasound-images-dataset (accessed on 20th February, 2023) 2. UDIAT: http://www2.docm.mmu.ac.uk/STAFF/M.Yap/dataset.php.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES