Abstract
In recent years, the growing importance of accurate semantic segmentation in ultrasound images has led to numerous advances in deep learning-based techniques. In this article, we introduce a novel hybrid network that synergistically combines convolutional neural networks (CNN) and Vision Transformers (ViT) for ultrasound image semantic segmentation. Our primary contribution is the incorporation of multi-scale CNN in both the encoder and decoder stages, enhancing feature learning capabilities across multiple scales. Further, the bottleneck of the network leverages the ViT to capture long-range high-dimension spatial dependencies, a critical factor often overlooked in conventional CNN-based approaches. We conducted extensive experiments using a public benchmark ultrasound nerve segmentation dataset. Our proposed method was benchmarked against 17 existing baseline methods, and the results underscored its superiority, as it outperformed all competing methods including a 4.6% improvement of Dice compared against TransUNet, 13.0% improvement of Dice against Attention UNet, 10.5% improvement of precision compared against UNet. This research offers significant potential for real-world applications in medical imaging, demonstrating the power of blending CNN and ViT in a unified framework.
Keywords: Ultrasound imaging, Vision transformer, Image semantic segmentation, Convolutional neural network
Introduction
Persistent postsurgical pain, defined as pain lasting more than 3–6 months after surgery (Merskey, 1986), presents a significant challenge in patient care. Management strategies for postsurgical pain typically encompass symptom control and disease modification (Kehlet, Jensen & Woolf, 2006). In clinical practice, the focus often shifts toward symptom control, which primarily relies on the use of narcotics and inhibitors (Baby & Jereesh, 2017). The frequent use of narcotics is associated with a range of unwanted side effects, including respiratory depression, nausea, vomiting, and other opioid-related adverse events (Bajwa & Haldar, 2015). Moreover, increased narcotic usage has been linked to extended hospital stays and heightened risk of depression (Armaghani et al., 2016). Some studies have suggested that indwelling catheters represent an alternative, safe, and effective method for postsurgical pain management (Wijayasinghe et al., 2016; Sola et al., 2012; Pacik, Nelson & Werner, 2008). However, the accurate placement of catheters is crucial, as incorrect placement can lead to unanticipated pain, opioid use, and potential complications, such as readmission or delayed hospital discharge (Hauritz et al., 2019). To address these challenges, various methods have been explored to enhance the precision of catheter placement and nerve location identification.
Nerve stimulation (NS) techniques have emerged to enhance the safety and precision of medical procedures, particularly in situations where traditional anatomic landmark techniques may lead to unintended punctures or cannulations (Pham-Dang et al., 2003; Kick et al., 1999; Copeland & Laxton, 2001). NS involves the stimulation of sensory nerves, inducing non-noxious sensations that effectively compete with and attenuate pain signals, thereby reducing pain perception. Additionally, NS has the potential to trigger the release of endorphins, natural pain-relieving chemicals, and modulate nerve activity implicated in pain signaling. Ultrasound technology has gained recognition as an alternative method to elevate the safety and quality of catheter placements in medical practice (Chan et al., 2003). Ultrasound techniques offer multifaceted advantages by enabling the visualization of nerve structures before injection, guiding the needle precisely to target nerves, and providing real-time visualization of the local anesthetic’s dispersion pattern. Numerous studies have demonstrated the superior performance of ultrasound-guided techniques over traditional anatomic landmarks and NS methods (Brass et al., 2015; Schnabel et al., 2013). In response to the ongoing pursuit of enhanced nerve identification and catheter placement precision, more advanced techniques have been proposed to further optimize these procedures.
Deep learning-based networks for ultrasound segmentation have emerged as the predominant choice, delivering remarkable segmentation performance at the pixel level. Convolutional neural networks (CNNs) have demonstrated the efficient capacity for extracting intricate features from grid-like data (Long, Shelhamer & Darrell, 2015; Ronneberger, Fischer & Brox, 2015; Huang et al., 2020; Oktay et al., 2018; Li et al., 2022). The UNet architecture revolutionized CNN-based segmentation with its symmetric encoder–decoder design, enabling impressive results even with limited datasets (Ronneberger, Fischer & Brox, 2015; Wang, Zhang & Voiculescu, 2021). However, CNNs have inherent limitations due to the localized nature of their convolutional operations, which can lead to under- or over-segmentation in complex ultrasound images. Addressing this challenge, the Attention UNet was introduced, showcasing its efficacy in handling variable small-sized organs by incorporating attention gates (AGs) within the UNet (Oktay et al., 2018). Further advancements by researchers like Chen, Yao & Zhang (2020) involved the fusion of the ResNet architecture (He et al., 2016) with attention mechanisms, thereby enhancing feature extraction and generating high-quality segmentation results for complex features. Another innovative approach, the AAUNet, adaptively selects receptive fields of varying scales from channel and spatial dimensions, leading to substantial improvements in breast lesion segmentation in ultrasound images (Chen et al., 2022).
Ultrasound images inherently incorporate non local features, resulting in ambiguous boundaries between target regions and backgrounds. Traditional UNet-based models face challenges in capturing long-range semantic dependencies within ultrasound images (Fang et al., 2023). Atrous or dilated convolution methods were introduced in these scenarios (Chen et al., 2017b; Chen et al., 2018). These methods expand the receptive field of convolutions without increasing the parameter count, allowing them to effectively aggregate multiscale contextual information. Atrous CNN have proven instrumental in achieving more accurate segmentation, particularly in scenarios involving intricate spatial structures and scales (Yu, Koltun & Funkhouser, 2017). For instance, Zhou, He & Jia (2020) applied atrous convolution to preserve resolution information in feature maps when segmenting brain tumor ultrasound images (Zhou, He & Jia, 2020), showcasing the versatility of these techniques in addressing segmentation challenges.
Recent advancements have seen the successful integration of CNN and transformer blocks to preserve global semantic information, with transformers demonstrating exceptional prowess in capturing intricate patterns and relationships in both natural language processing (Devlin et al., 2018; Vaswani et al., 2023) and computer vision domains (Parmar et al., 2018; Liu et al., 2021). A novel adaptation of the Transformer for computer vision, known as the ViT, eliminates the need for convolutions to extract features from images (Dosovitskiy et al., 2020). The ViT segments images into discrete non-overlapping patches. Spatial positioning information is then introduced to these patches through position encodings, and they are subsequently passed through standard transformer layers. This allows the ViT to effectively model both local and global semantic dependencies. Further augmenting this progress, the Segformer incorporates a Bilinear Fusion mechanism to efficiently merge multi-level feature maps, enhancing both receptive field and resolution for optimized segmentation results (Xie et al., 2021). The TransUNet offers a compelling solution with remarkable segmentation performance, effectively marrying high-resolution spatial details from CNN features with the contextual breadth of transformers to address inherent locality limitations and mitigate feature resolution loss, typically associated with pure transformers (Chen et al., 2021). The Swin-UNet, by combining a symmetric encoder–decoder structure with skip connections and integrating local-to-global self-attention, marks a significant advancement in image segmentation, optimizing transformer computations and enhancing segmentation efficiency (Cao et al., 2022). Lin and collaborators have integrated Swin Transformers and Multi-scale Vision Transformers (Chen, Fan & Panda, 2021) into the UNet, fostering excellent long-range dependencies between features of different scales (Lin et al., 2022). Additionally, CSwin-PNet was proposed to further enhance long-range dependency modeling, particularly tailored for ultrasound breast segmentation (Yang & Yang, 2023).
Given the considerations highlighted, we recognized the significance of global modeling within CNN. Drawing inspiration from TransUNet and atrous convolutions, we introduce our novel approach, referred to as the Modified CNN & ViT hybrid Encoder-Decoder segmentation network with multi-scale information fusion approach (MCV-UNet). To our knowledge, this marks the first endeavor to integrate CNN and ViT explicitly for ultrasound nerve segmentation. Our contributions in this study can be delineated as follows:
-
1.
Inspired by the burgeoning success of the Vision Transformer in the domain of computer vision, we have further integrated the ViT-layer within the Encoder-Decoder segmentation paradigm, enhancing its feature extraction prowess.
-
2.
Recognizing the importance of capturing intricate details that span from local nuances to broader patterns, we introduce various atrous CNN layers. These layers augment the network’s receptive field, bolstering its ability to discern and process multi-scale spatial hierarchies.
-
3.
To validate the efficacy of our approach, we compared MCV-UNet against an array of established baseline methods. The empirical evaluations underscored our network’s superior capabilities on a public dataset, yielding competitive results against 15 baseline methods.
The remainder of this article is structured as follows: ‘Related work’ reviews the relevant literature, highlighting key developments in CNN and ViT utilized to medical image segmentation. ‘Approach’ details the proposed approach, MCV-UNet, including the network framework, analytical techniques, and related equations. ‘Results’ discusses the results obtained with MCV-UNet, covering data sources, implementation details, and evaluation criteria. It also provides an in-depth discussion of these results from different perspectives. ‘Conclusion’ is the conclusion including remarks, summarizing the superior performance of MCV-UNet and suggesting ideas for future research in this field. To aid in the clarity and readability of this article, a table of abbreviations is provided in Table 1.
Table 1. Abbreviation instructions.
Abbreviation | Full form |
---|---|
CNN | Convolutional neural networks |
ViT | Vision transformers |
NS | Nerve stimulation |
AGs | Attention gates |
FCN | Fully convolutional network |
Dice | Dice coefficient |
Acc | Accuracy |
Pre | Precision |
Sen | Sensitivity |
Spec | Specificity |
Cost | Computational cost |
LN | Layer normalization |
MSA | Multi-head self-attention |
MLP | Multilayer perceptron |
TP | True positive |
FP | False positive |
TN | True negative |
FN | False negative |
Related Work
Medical image segmentation with CNN
CNN has initially emerged as the predominant methods for image processing tasks (Milletari, Navab & Ahmadi, 2016; Chen et al., 2018; Lv et al., 2020; Ali, Qureshi & Shah, 2023). In the domain of medical image processing, where the desired output extends beyond a single class label, the need for precise segmentation of organs or tumors is paramount. Pioneering efforts by Cireşan, Meier & Schmidhuber (2012) leveraged deep neural network networks trained on GPUs, leading to substantial improvements in recognition rates on medical image datasets. The introduction of the fully convolutional network (FCN) marked a crucial development by striking a balance between capturing global and local information through the integration of multi-resolution layers (Long, Shelhamer & Darrell, 2015). To further enhance training efficiency with limited data, Ronneberger introduced the symmetric UNet architecture, extending the contracting network by incorporating successive layers with skip connections (Ronneberger, Fischer & Brox, 2015). UNet quickly gained popularity for its remarkable ability to learn invariance from medical images. LinkNet innovatively directly linked the encoder to the corresponding decoder, ensuring precise predictions without compromising network processing speed (Chaurasia & Culurciello, 2017). Subsequent advancements in UNet-based networks, like the Attention UNet with its AGs mechanisms, enabled networks to focus on targets of complex shape and size, expanding its applications in ultrasound image segmentation (Oktay et al., 2018). Res-UNet was specifically designed for ultrasound nerve segmentation, enhancing accuracy through the incorporation of dense atrous convolutions and residual multiple posing modules compared to the traditional UNet (Wang, Shen & Zhou, 2019). Furthermore, researchers have explored combining recurrent neural networks with residual neural networks to achieve improved organ segmentation performance (Alom et al., 2018). Addressing the growing demand for precise medical image segmentation, UNet3+ maximized feature map utilization through full-scale connections (Huang et al., 2020). Transfer Learning techniques were also incorporated with the UNet architecture, as demonstrated by Cheng & Lam (2021) who applied their network successfully to lung ultrasound segmentation, leveraging mechanisms for detecting edges, shapes, and textures from ultrasound images. Additionally, the Dense-PSP-UNet introduced an innovative Pyramid Scene Parsing (PSP) module, surpassing skip connection settings in performance and employing Contrast Limited Adaptive Histogram Equalization (CLAHE) (Reza, 2004) to reduce image noise levels during training (Ansari et al., 2023).
Medical image segmentation with transformers
The transformer architecture, initially pivotal in sequential processing, marked a paradigm shift with its self-attention mechanism, enabling unprecedented performance in various classification tasks (Vaswani et al., 2017). In computer vision, the ViT replaced traditional convolutional layers with a novel approach of segmenting images into non-overlapping patches, treated as linear embeddings. This method facilitated contextual relationships between patches through self-attention, enhancing the network’s comprehension of the entire image (Dosovitskiy et al., 2020; Wang, Zhao & Ni, 2022; Chen et al., 2021; Liu, Hu & Chen, 2023). ViT has set new benchmarks in object detection (Fang et al., 2021), rivaling state-of-the-art CNN architectures, particularly when pre-trained on extensive datasets (Dosovitskiy et al., 2020). The introduction of axial (Ho et al., 2019) and hierarchical attention (Yang et al., 2016) further refined ViT, enabling more precise segmentation. A significant advantage of ViT is its capacity to handle varied image sizes, crucial for intricate ultrasound images. In medical image segmentation, where precision is critical for diagnosis and treatment, traditional methods face challenges like varying contrasts and subtle pathological indicators. TransUNet combined CNN’s local detail capture with transformers’ holistic view, enhancing the understanding of medical images (Chen et al., 2021). Swin-Unet, leveraging transformer blocks, adeptly handles high-resolution medical scans (Liu et al., 2021; Cao et al., 2022). The hybrid CNN-Transformer network further innovated by integrating large-kernel convolution, effectively capturing multi-scale information (Liu, Hu & Chen, 2023). The ViT-Patch introduced a secondary task on the patch tokens, in addition to the primary task on the class token, demonstrating superior performance compared to the standard ViT in breast ultrasound segmentation. token (Feng et al., 2023). HA-UNet’s introduction of local–global transformer blocks represents a significant step in reducing computational complexity without sacrificing segmentation efficiency (Zhang et al., 2024). The inclusion of a cross attention block in HA-UNet not only improved feature integration but also demonstrated significant advancements in ultrasound breast lesion segmentation.
Approach
Architecture overview
In the domain of deep learning applied to image segmentation, the objective is to map an input image x to its segmented inference y. This mapping is denoted as ypred = f(x; θ), where f is the deep learning network, θ represents the network’s parameters, and ypred is the predicted segmentation of each pixel, where pred ∈ [0, 1]. The corresponding ground truth for the input image x is represented as ygt. During the training phase, we use a dataset consisting of batches of paired data represented as (x, ygt) ∈ Dtrain. Our primary aim during training is to optimize the parameters θ to minimize the difference between ypred and ygt. For evaluation on unseen data, we use (x, ygt) ∈ Dtest and assess the network’s performance by comparing ypred to ygt. The MCV-UNet, a novel approach in medical image segmentation for ultrasound images, is depicted in Fig. 1. The architecture, which integrates CNN and ViT, consists of an encoder, bottleneck, decoder, and skip connections. Built upon the foundational UNet structure (Ronneberger, Fischer & Brox, 2015), MCV-UNet innovates with key components atrous convolutional and ViT layers. The process begins with two 3 × 3 atrous convolutional layers in the encoder, designed to extract multi-scale features while expanding the network’s receptive field without significantly increasing computations (Chen et al., 2017a). This is followed by standard convolution-based encoders and max-pooling layers, effectively balancing spatial dimension reduction and computational efficiency. In the symmetric design, the feature maps from the encoder aid each upsampling step in the decoder. A 2 × 2 deconvolution layer halves the feature channels, and these are merged with corresponding encoder feature maps via skip connections, preserving crucial information.
We introduce a bottleneck with two transformer blocks, leveraging ViT for its segmentation accuracy (Dosovitskiy et al., 2020). This unique combination of atrous convolution and ViT allows MCV-UNet to capture both local details and global context effectively, a crucial requirement in medical image segmentation. The final expanding layer in the decoder then maps feature vectors to the desired class numbers, ensuring the output matches the input resolution. MCV-UNet’s design is a strategic evolution from conventional UNet, inspired by TransUNet’s hybrid CNN-Transformer approach (Chen et al., 2021). The specific functionalities of ViTs and atrous convolutions, crucial to MCV-UNet’s performance, are further detailed in the subsequent sections.
Vision transformer layer
Two successive Vision Transformer blocks serves as a key element in the bottleneck between the encoder and decoder is illustrated in Fig. 2. Within each Vision Transformer block, we applied a layer normalization (LN), multi-head self-attention (MSA), a two-layer multilayer perceptron (MLP) with GELU (Ba, Kiros & Hinton, 2016; Hendrycks & Gimpel, 2016). A residual connection was applied each module (He et al., 2016). The computation within these continuous Transformer blocks is illustrated as follows:
(1) |
(2) |
(3) |
(4) |
here, and zl represent the outputs of the MSA module and the MLP module of the lth block, respectively.
Drawing inspiration from previous works (Pan et al., 2022; Keles, Wijewardena & Hegde, 2023; Wang & Ma, 2023; Qin et al., 2022; Zhang et al., 2023), our self-attention computation strategy adheres to the principles of scaled dot-product attention (Vaswani et al., 2023). This approach leverages the efficiency of dot-product attention, optimizing the use of matrix multiplication (Bahdanau, Cho & Bengio, 2014). The self-attention computation can be illustrated as:
(5) |
where Q, K, and V are matrices representing queries, keys, and values, respectively, with dimensions Q, K, V ∈ ℝM2×d, where M2 denotes the number of patches in a window and d represents the query or key dimension. The bias term B is derived from the bias matrix , with .
Atrous convolution layer
In the classical encoder–decoder architecture, the repeated operations of max-pooling and striding at consecutive layers often lead to a substantial reduction in the spatial resolution of the resulting feature maps. While skip connections and deconvolutional layers can help recover some of this lost spatial information, MCV-UNet takes a further step to mitigate this issue by incorporating atrous convolution within the encoder–decoder architecture.
Atrous convolution, initially introduced for the computation of the undecorated wavelet transform in the “algorithme à trous” scheme (Holschneider et al., 1990), has demonstrated high performance in various applications, including semantic segmentation (Chen et al., 2017a). In the context of two-dimensional data, atrous convolution is defined as follows:
(6) |
here, the rate parameter r corresponds to the stride of the sampled input signal, x[i] represents the two-dimensional input signal, and the atrous rate r convolves the input signal with the upsampled filter w[k] by introducing r − 1 zeros between consecutive filter values along each spatial dimension (Chen et al., 2017b; Gu et al., 2019). The parameter K denotes the length of the filter w[k], and the standard convolution corresponds to the special case of an atrous rate r = 1. Adjusting the atrous rate provides the network with flexibility in terms of the field of view and enables the generation of larger outputs without significantly increasing computational demands (Chen et al., 2017a). Previous works have incorporated atrous convolution in various ways, including within encoder–decoder blocks (Chen et al., 2018; Chen et al., 2017b), skip connections (Wang & Voiculescu, 2021), and the module bridging the encoding and decoding stages to extract dense features (Lv et al., 2020; Gu et al., 2019; Pan et al., 2019; Ma, Gu & Wang, 2024).
In the architecture of MCV-UNet, inspired by the encoder–decoder structure with atrous convolution (Chen et al., 2018), we placed batch normalization layers before atrous convolution operations with atrous rates r = 3 in the encoder blocks and r = 1 (standard convolution) in the decoder blocks, as illustrated in Fig. 3. The introduction of holes (with r = 3) in the down-sampling process facilitates the computation of responses at all image positions while introducing zeros between filter values. This increases the size of the filter compared to the standard convolution layer, but computations only consider the values of non-zero filter elements, ensuring a constant number of filter parameters and computational operations. Overall, this approach offers the advantage of controlling feature resolution, enhancing the receptive field of the network without sacrificing image resolution.
Results
Dataset
In our experiments, we employed the Nerve Segmentation database, a publicly available resource from the Kaggle Competition platform (Anna Montoya et al., 2016). This database is integral to medical imaging, particularly for the analysis of the brachial plexus nerve, a critical area often studied in ultrasound imaging. This dataset comprises 5,640 256 × 256 ultrasound images, distinctly split into 1,128 testing and 4,512 validation samples, with no overlap between training, validation, and testing sets. Each image encompasses a 2D ultrasound scan of the nerve alongside a meticulously manually annotated 2D segmentation mask, serving as ground truth. The images present a unique challenge due to the random distribution of the nerve within them, demanding precise segmentation skills. To facilitate uniform analysis, all images underwent normalization, scaling pixel values to the range [0, 1], thereby simplifying the task of differentiating the nerve from surrounding tissues. This approach ensures accurate segmentation by leveraging expert annotations and standardized image processing techniques.
Implementation details
The implementation of our approach was developed using Python 3 and TensorFlow (Abadi et al., 2015). Our experiments were conducted on a robust computing setup, featuring an Intel Xeon CPU with 2 vCPUs and 13 GB of RAM, and significantly accelerated with an NVIDIA A100 GPU, equipped with 40 GB of VRAM, known for its high computational efficiency in deep learning tasks. We adapted several networks from established sources, specifically segmentation models and Keras-UNet-Collections, applying necessary modifications to optimize them for our specific dataset. These adaptations were crucial in handling our dataset’s unique characteristics.
During the training phase, consistency in parameters across all networks was maintained to ensure fair comparative analysis. We employed the Adam Optimizer (Kingma & Ba, 2014), renowned for its efficiency in computing gradients, setting the learning rate to to 10−4, batch size to 8, and the number of training epochs to 50. Batch normalization layers (Ioffe & Szegedy, 2015) were strategically incorporated to enhance training speed and stability. The details of the hyper-parameter setting in the experiment is illustrated in Table 2.
Table 2. The hyper-parameter setting for MCV-UNet and all baseline methods.
Epoch | Optimizer | Learning rate | Batch size | Dataset |
---|---|---|---|---|
50 | Adam | 10−4 | 8 | 5640 × 256 × 256 |
The network’s performance was evaluated using the Dice coefficient-based loss, a standard metric in image segmentation tasks, which quantifies the similarity between predicted and ground truth segmentation. We saved the network from the epoch showing the best performance for testing phase segmentation. On our specified hardware, training a single network typically required 3 to 5 h, depending on the network’s complexity and architecture.
Metrics
To comprehensively evaluate the performance of MCV-UNet, a diverse set of evaluation metrics are utilized. These metrics encompass a range of criteria, including the Dice coefficient (Dice), Accuracy (Acc), Precision (Pre), Sensitivity (Sen), Specificity (Spec), and the parameters of network as computational cost metrics (Cost). Each of these metrics offers a unique perspective on the effectiveness of MCV-UNet in segmenting medical images. The details of our evaluation metrics can be outlined as follows:
(7) |
(8) |
(9) |
(10) |
(11) |
where, TP represents the number of true positives, TN denotes the number of true negatives, FP signifies the number of false positives, and FN stands for the number of false negatives.
By employing these diverse metrics, a comprehensive assessment of MCV-UNet’s segmentation performance could be achieved. Each metric contributes valuable insights into different aspects of the network’s performance, enabling us to evaluate the effectiveness and accuracy of MCV-UNet in the context of medical image segmentation.
Comparison with state-of-the-arts
To evaluate the performance of MCV-UNet in the context of medical image segmentation, we conducted an extensive comparison with 17 baseline networks, including a diverse range of architectural designs, each with its own strengths and characteristics. To demonstrate the effectiveness of MCV-UNet, we use ViT block bridge the encoder and decoder, with comparisons made against classical CNN networks and their variants. Furthermore, we evaluated the impact of the encoder and decoder design, incorporating atrous convolution layers, by contrasting the result with existing hybrid CNN-ViT networks. The networks compared include: UNet (Ronneberger, Fischer & Brox, 2015), UNet-ResNet34 (Ronneberger, Fischer & Brox, 2015), UNet-MobileNet (Ronneberger, Fischer & Brox, 2015), UNet-InceptionV3 (Ronneberger, Fischer & Brox, 2015), Linknet (Chaurasia & Culurciello, 2017), Linknet-MobileNet (Chaurasia & Culurciello, 2017), FPN-ResNet34 (Lin et al., 2017), FPN-MobileNet (Lin et al., 2017), TransUNet (Chen et al., 2021), FPN-InceptionV3 (Lin et al., 2017), VNet (Milletari, Navab & Ahmadi, 2016), AttentionUNet (Oktay et al., 2018), UNet3+ (Huang et al., 2020), U2-Net (Qin et al., 2020), RARUNet (Wang, Zhang & Voiculescu, 2021), QAPNet (Wang & Voiculescu, 2021), and R2UNet (Alom et al., 2018). This comprehensive comparison allows us to demonstrate the unique strengths and capabilities of MCV-UNet in the context of medical image segmentation.
Qualitative results
Figure 4 displays qualitative comparison results with three randomly chosen ultrasound medical images alongside their corresponding ground truths. The original raw images are removed due to Kaggle Policy. For each example, predictions generated by 17 baseline methods are compared with the results from MCV-UNet. The results of these visual analyses yield valuable insights into the performance of each approach. Classical CNN-based methods, including UNet (Ronneberger, Fischer & Brox, 2015), Attention UNet (Oktay et al., 2018), and V-Net (Milletari, Navab & Ahmadi, 2016), tend to exhibit issues of over-segmentation or under-segmentation. For instance, in the first example, UNet-MobileNet over-segments the nerve while V-Net under-segments it. This observation underscores the superior capability of MCV-UNet in effectively encoding global contexts and distinguishing the semantics. In addition, in the context of existing hybrid CNN-ViT networks, the predictions generated by TransUNet (Chen et al., 2021) demonstrate coarser characteristics than those by MCV-UNet, particularly with regard to boundary and shape. In the third example, MCV-UNet displays excellent alignment with nerve boundary of the ground truth, whereas TransUNet predicts more false positives. These visual comparisons proves the superior performance of our network, characterized by its capacity to preserve detailed shape information, resulting in fewer false positives and false negatives compared to the baseline methods. This superiority is attributed to the successful combination of CNN and ViT architectures in preserving high-level global information and low-level details, while minimizing spatial information loss with atrous convolution layers.
Quantitative results
Table 3 presents a comprehensive quantitative evaluation of our MCV-UNet in comparison to the 17 baseline methods in Tables 3, 4, 5. The quantitative results proved the exceptional performance of MCV-UNet on most evaluation metrics. For the main criterion metric of dice coefficient, MCV-UNet achieves a remarkable result of 62.51%, surpassing the second-ranked network by 0.47%. In terms of accuracy and precision, MCV-UNet outperforms all competitors, with a 0.18% increase in Acc and a 1.16% increase in Pre compared to the second-best network. MCV-UNet exhibits competitive performance in sensitivity and specificity metrics, aligning with the top-performing networks. Regarding computational cost, MCV-UNet falls within the median range of all trained networks, showing a slight advantage over our derived architecture, TransUNet. These quantitative results indicate MCV-UNet’s competitive edge across multiple evaluation metrics relative to the 17 baseline networks. The parameters of MCV-UNet is 43%, 76&, 49% lower than UNet with mobilenet as network backbone, LinkNet, and QAPNet. Notably, it is also slightly lower than the current advanced ViT-based TransUNet due to the modified atrous CNN is utilized. They also emphasize capabilities of MCV-UNet of combining the strengths of classical CNN and hybrid CNN-ViT networks.
Table 3. The performance of MCVUNet with other baseline methods on ultrasound nerve segmentation test set.
Network | Dice | Acc | Pre | Sen | Spec | Cost |
---|---|---|---|---|---|---|
UNet | 0.6217 | 0.9910 | 0.6179 | 0.6256 | 0.9953 | 1,967,041 |
UNet-ResNet34 | 0.6165 | 0.9910 | 0.6218 | 0.6130 | 0.9955 | 24,456,160 |
UNet-MobileNet | 0.6184 | 0.9907 | 0.6038 | 0.6349 | 0.9950 | 8,336,343 |
UNet-InceptionV3 | 0.6173 | 0.9899 | 0.5599 | 0.6997 | 0.9934 | 29,933,111 |
LinkNet | 0.6218 | 0.9911 | 0.6286 | 0.6168 | 0.9956 | 20,325,137 |
LinkNet-MobileNet | 0.6067 | 0.9896 | 0.5474 | 0.6840 | 0.9932 | 4,546,071 |
FPN-ResNet34 | 0.6172 | 0.9906 | 0.5941 | 0.6435 | 0.9947 | 23,930,960 |
FPN-MobileNet | 0.5971 | 0.9907 | 0.6099 | 0.5859 | 0.9955 | 6,103,111 |
TransUNet | 0.5976 | 0.9888 | 0.5193 | 0.7037 | 0.9922 | 4,675,335 |
FPN-InceptionV3 | 0.6222 | 0.9904 | 0.5803 | 0.6745 | 0.9942 | 25,029,287 |
VNet | 0.5955 | 0.9896 | 0.5533 | 0.6491 | 0.9937 | 3,690,129 |
AttentionUNet | 0.5533 | 0.9887 | 0.5184 | 0.5934 | 0.9934 | 638,322 |
UNet3+ | 0.5890 | 0.9899 | 0.5695 | 0.6099 | 0.9945 | 497,848 |
U2-Net | 0.4366 | 0.9892 | 0.5738 | 0.3523 | 0.9969 | 3,775,677 |
R2UNet | 0.6115 | 0.9910 | 0.6230 | 0.6004 | 0.9956 | 1,448,215 |
RARUNet | 0.6219 | 0.9910 | 0.6181 | 0.6257 | 0.9953 | 11,793,638 |
QAPNet | 0.6218 | 0.9910 | 0.6180 | 0.6257 | 0.9954 | 9,472,052 |
Ours | 0.6251 | 0.9928 | 0.6359 | 0.6912 | 0.9960 | 4,675,329 |
Notes.
The best performance results are highlighted in bold. The second-best performance of MCV-UNet is highlighed with an underline.
Table 4. The sensitivity analysis of atrous CNN setting.
Dilation rate | 1 | 2 | 3 | 4 | 5 | 6 |
---|---|---|---|---|---|---|
Dice | 0.5976 | 0.6105 | 0.6251 | 0.6248 | 0.6296 | 0.6175 |
Acc | 0.9888 | 0.9894 | 0.9928 | 0.9912 | 0.9907 | 0.9910 |
Pre | 0.5193 | 0.5423 | 0.6359 | 0.6323 | 0.5955 | 0.6186 |
Spec | 0.9922 | 0.9929 | 0.9960 | 0.9957 | 0.9946 | 0.9954 |
Notes.
The best performance results are highlighted in bold. The second-best performance of MCV-UNet is highlighed with an underline.
Table 5. The ablation study of MCV-UNet on ultrasound nerve segmentation test set.
Multi-Scale CNN | Self-Attention | Dice | Acc | Pre | Sen | Spec |
---|---|---|---|---|---|---|
0.6217 | 0.9910 | 0.6179 | 0.6256 | 0.9953 | ||
✓ | 0.6220 | 0.9903 | 0.5780 | 0.6706 | 0.9941 | |
✓ | 0.5976 | 0.9888 | 0.5193 | 0.7037 | 0.9922 | |
✓ | ✓ | 0.6251 | 0.9928 | 0.6359 | 0.6912 | 0.9960 |
Notes.
The best performance results are highlighted in bold.
Sensitivity analysis & ablation study
In addition to our primary experiments, we carried out a sensitivity analysis focused on the hyper-parameter setting related to the dilated rate in our multi-scale CNN. This analysis is essential to determine the optimal configuration for effectively segmenting nerves in ultrasound images. The detailed results of this analysis are presented in Table 4. These findings validate the effectiveness of the dilated rate settings in our proposed MCV-UNet.
To study the individual and collective impact of the multi-scale modules proposed in our network, we conducted an ablation study, the results of which are detailed in Table 5. This study methodically explores the effects of omitting or modifying various components of our network. These findings not only validate the efficacy of each proposed contribution but also highlight their synergistic effect in enhancing the accuracy and robustness of the ultrasound nerve segmentation process.
Conclusion
In this study, we studied the combination of modified CNN and ViT to address the intricate challenge of ultrasound nerve segmentation. Recognizing the limitations inherent in conventional CNN-based networks, especially their restricted capacity to exploit long-range semantic dependencies in ultrasound images, we proposed the MCV-UNet. This novel design represents a modified encoder–decoder framework that seamlessly combines the robust capabilities of both CNN and ViT, while it integrates atrous convolution layers to effectively recover lost spatial information. This kind of multi-scale feature information extraction is valuable in ultrasound imaging, because the nerve structure is complex, and the location, size of nerve can be different in each of ultrasound image. Considering other modalities images, such as CT, MRI, and PET, the MCV-UNet is also valuable to be explored especially when the region of interest (ROI) is complex and should be recognized based on both of the local- and global-based features.
The qualitative and quantitative evaluations indicates that the proposed network outperformed 17 classical baseline methods, exhibiting fewer FP and FN—a testament to its robustness and precision. The integration of various dilated CNN layers further amplified its feature extraction capabilities, bridging the gap between local and global contextual understanding in the images.
Future work might consider refining the network architecture, introducing novel attention mechanisms, or expanding the approach to other challenging medical imaging domains. The computational burden should also be further decreased, because the current parameters of network is still high due to the utilization of ViT. Different types of CNN, network pruning, and knowledge distillation can also be studied to enable the efficiency of the network.
Supplemental Information
Funding Statement
The authors received no funding for this work.
Additional Information and Declarations
Competing Interests
The authors declare there are no competing interests.
Author Contributions
Zihong Xu conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.
Ziyang Wang conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, authored or reviewed drafts of the article, and approved the final draft.
Data Availability
The following information was supplied regarding data availability:
The code is available in the Supplemental File.
The raw data is publicly available at Kaggle: https://www.kaggle.com/competitions/ultrasound-nerve-segmentation.
References
- Abadi et al. (2015).Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jia Y, Jozefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X. TensorFlow: large-scale machine learning on heterogeneous systems. https://www.tensorflow.org 2015
- Ali, Qureshi & Shah (2023).Ali H, Qureshi R, Shah Z. Artificial intelligence–based methods for integrating local and global features for brain cancer imaging: scoping review. JMIR Medical Informatics. 2023;11(1):e47445. doi: 10.2196/47445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alom et al. (2018).Alom MZ, Hasan M, Yakopcic C, Taha TM, Asari VK. Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation. 2018 doi: 10.48550/arXiv.1802.06955.1802.06955 [DOI]
- Ansari et al. (2023).Ansari MY, Yang Y, Meher PK, Dakua SP. Dense-PSP-UNet: a neural network for fast inference liver ultrasound segmentation. Computers in Biology and Medicine. 2023;153:106478. doi: 10.1016/j.compbiomed.2022.106478. [DOI] [PubMed] [Google Scholar]
- Anna Montoya et al. (2016).Anna Montoya H, kaggle446. shirzad. Cukierski W, yffud Ultrasound Nerve Segmentation. Kaggle. 2016. https://kaggle.com/competitions/ultrasound-nerve-segmentation https://kaggle.com/competitions/ultrasound-nerve-segmentation
- Armaghani et al. (2016).Armaghani SJ, Lee DS, Bible JE, Shau DN, Kay H, Zhang C, McGirt MJ, Devin CJ. Increased preoperative narcotic use and its association with postoperative complications and length of hospital stay in patients undergoing spine surgery. Journal of Spinal Disorders and Techniques. 2016;29(2):e93–e98. doi: 10.1097/BSD.0000000000000109. [DOI] [PubMed] [Google Scholar]
- Ba, Kiros & Hinton (2016).Ba JL, Kiros JR, Hinton GE. Layer normalization. 20161607.06450
- Baby & Jereesh (2017).Baby M, Jereesh A. Automatic nerve segmentation of ultrasound images. 2017 International conference of electronics, communication and aerospace technology (ICECA), volume 1; Piscataway. 2017. pp. 107–112. [Google Scholar]
- Bahdanau, Cho & Bengio (2014).Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. 20141409.0473
- Bajwa & Haldar (2015).Bajwa SJS, Haldar R. Pain management following spinal surgeries: an appraisal of the available options. Journal of Craniovertebral Junction & Spine. 2015;6(3):105. doi: 10.4103/0974-8237.161589. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brass et al. (2015).Brass P, Hellmich M, Kolodziej L, Schick G, Smith AF. Ultrasound guidance versus anatomical landmarks for internal jugular vein catheterization. Cochrane Database of Systematic Reviews. 2015;1(1):CD006962. doi: 10.1002/14651858.CD006962.pub2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cao et al. (2022).Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M. Swin-Unet: unet-like pure transformer for medical image segmentation. European conference on computer vision; Cham. 2022. [Google Scholar]
- Chan et al. (2003).Chan VW, Perlas A, Rawson R, Odukoya O. Ultrasound-guided supraclavicular brachial plexus block. Anesthesia & Analgesia. 2003;97(5):1514–1517. doi: 10.1213/01.ANE.0000062519.61520.14. [DOI] [PubMed] [Google Scholar]
- Chaurasia & Culurciello (2017).Chaurasia A, Culurciello E. Linknet: exploiting encoder representations for efficient semantic segmentation. 2017 IEEE visual communications and image processing (VCIP); Piscataway. 2017. pp. 1–4. [Google Scholar]
- Chen, Fan & Panda (2021).Chen C-FR, Fan Q, Panda R. Crossvit: cross-attention multi-scale vision transformer for image classification. Proceedings of the IEEE/CVF international conference on computer vision; Piscataway. 2021. pp. 357–366. [Google Scholar]
- Chen et al. (2022).Chen G, Li L, Dai Y, Zhang J, Yap MH. AAU-net: an adaptive attention U-net for breast lesions segmentation in ultrasound images. IEEE Transactions on Medical Imaging. 2022;42(5):1289–1300. doi: 10.1109/TMI.2022.3226268. [DOI] [PubMed] [Google Scholar]
- Chen et al. (2021).Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, Lu L, Yuille AL, Zhou Y. Transunet: transformers make strong encoders for medical image segmentation. 20212102.04306
- Chen et al. (2017a).Chen L-C, Papandreou G, Kokkinos I, Murphy K, Yuille AL. Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2017a;40(4):834–848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27. [DOI] [PubMed] [Google Scholar]
- Chen et al. (2017b).Chen L-C, Papandreou G, Schroff F, Adam H. Rethinking atrous convolution for semantic image segmentation. 2017b1706.05587
- Chen et al. (2018).Chen L-C, Zhu Y, Papandreou G, Schroff F, Adam H. Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of the European conference on computer vision (ECCV); 2018. pp. 801–818. [Google Scholar]
- Chen, Yao & Zhang (2020).Chen X, Yao L, Zhang Y. Residual attention u-net for automated multi-class segmentation of COVID-19 chest CT images. 20202004.05645
- Cheng & Lam (2021).Cheng D, Lam EY. Transfer learning U-Net deep learning for lung ultrasound segmentation. 20212110.02196
- Cireşan, Meier & Schmidhuber (2012).Cireşan D, Meier U, Schmidhuber J. Multi-column deep neural networks for image classification. 2012 IEEE conference on computer vision and pattern recognition; Piscataway. 2012. [Google Scholar]
- Copeland & Laxton (2001).Copeland SJ, Laxton MA. A new stimulating catheter for continuous peripheral nerve blocks. Regional Anesthesia and Pain Medicine. 2001;26(6):589. doi: 10.1053/rapm.2001.26215. [DOI] [PubMed] [Google Scholar]
- Devlin et al. (2018).Devlin J, Chang M.-W., Lee K, Toutanova K. Bert: pre-training of deep bidirectional transformers for language understanding. 20181810.04805
- Dosovitskiy et al. (2020).Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S. An image is worth 16x16 words: transformers for image recognition at scale. 20202010.11929
- Dosovitskiy et al. (2020).Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M. An image is worth 16x16 words: transformers for image recognition at scale. 20202010.11929
- Fang et al. (2021).Fang Y, Liao B, Wang X, Fang J, Qi J, Wu R, Niu J, Liu W. You only look at one sequence: rethinking transformer in vision through object detection. Advances in Neural Information Processing Systems. 2021;34:26183–26197. [Google Scholar]
- Fang et al. (2023).Fang C, Wang Q, Cheng L, Gao Z, Pan C, Cao Z, Zheng Z, Zhang D. Reliable mutual distillation for medical image segmentation under imperfect annotations. IEEE Transactions on Medical Imaging. 2023;42(6):1720–1734. doi: 10.1109/TMI.2023.3237183. [DOI] [PubMed] [Google Scholar]
- Feng et al. (2023).Feng H, Yang B, Wang J, Liu M., Yin L, Zheng W, Yin Z, Liu C. Identifying malignant breast ultrasound images using ViT-patch. Applied Sciences. 2023;13(6):3489. doi: 10.3390/app13063489. [DOI] [Google Scholar]
- Gu et al. (2019).Gu Z, Cheng J, Fu H, Zhou K., Hao H, Zhao Y, Zhang T, Gao S, Liu J. Ce-net: context encoder network for 2d medical image segmentation. IEEE Transactions on Medical Imaging. 2019;38(10):2281–2292. doi: 10.1109/TMI.2019.2903562. [DOI] [PubMed] [Google Scholar]
- Hauritz et al. (2019).Hauritz RW, Hannig KE, Balocco AL, Peeters G, Hadzic A, Børglum J, Bendtsen TF. Peripheral nerve catheters: a critical review of the efficacy. Best Practice & Research Clinical Anaesthesiology. 2019;33(3):325–339. doi: 10.1016/j.bpa.2019.07.015. [DOI] [PubMed] [Google Scholar]
- He et al. (2016).He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition; Piscataway. 2016. pp. 770–778. [Google Scholar]
- Hendrycks & Gimpel (2016).Hendrycks D, Gimpel K. Gaussian error linear units (gelus) 20161606.08415
- Ho et al. (2019).Ho J, Kalchbrenner N, Weissenborn D, Salimans T. Axial attention in multidimensional transformers. 20191912.12180
- Holschneider et al. (1990).Holschneider M, Kronland-Martinet R, Morlet J, Tchamitchian P. A real-time algorithm for signal analysis with the help of the wavelet transform. In: Combes JM, Grossmann A, Tchamitchian P, editors. Wavelets. Inverse problems and theoretical imaging; Berlin, Heidelberg. 1990. pp. 286–297. [DOI] [Google Scholar]
- Huang et al. (2020).Huang H, Lin L, Tong R, Hu H, Zhang Q, Iwamoto Y, Han X, Chen Y-W, Wu J. Unet 3+: a full-scale connected unet for medical image segmentation. ICASSP 2020-2020 IEEE international conference on acoustics, speech and signal processing (ICASSP); Piscataway. 2020. pp. 1055–1059. [Google Scholar]
- Ioffe & Szegedy (2015).Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. International conference on machine learning.2015. [Google Scholar]
- Kehlet, Jensen & Woolf (2006).Kehlet H, Jensen TS, Woolf CJ. Persistent postsurgical pain: risk factors and prevention. The Lancet. 2006;367(9522):1618–1625. doi: 10.1016/S0140-6736(06)68700-X. [DOI] [PubMed] [Google Scholar]
- Keles, Wijewardena & Hegde (2023).Keles FD, Wijewardena PM, Hegde C. On the computational complexity of self-attention. International conference on algorithmic learning theory.2023. [Google Scholar]
- Kick et al. (1999).Kick O, Blanche E, Pham-Dang C, Pinaud M, Estebe JP. A new stimulating stylet for immediate control of catheter tip position in continuous peripheral nerve blocks. Anesthesia & Analgesia. 1999;89(2):533–534. doi: 10.1213/00000539-199908000-00062. [DOI] [PubMed] [Google Scholar]
- Kingma & Ba (2014).Kingma DP, Ba J. Adam: a method for stochastic optimization. 20141412.6980
- Li et al. (2022).Li W, Tang YM, Wang Z, Yu KM, To S. Atrous residual interconnected encoder to attention decoder framework for vertebrae segmentation via 3D volumetric CT images. Engineering Applications of Artificial Intelligence. 2022;114:105102. doi: 10.1016/j.engappai.2022.105102. [DOI] [Google Scholar]
- Lin et al. (2022).Lin A, Chen B, Xu J, Zhang Z, Lu G, Zhang D. Ds-transunet: dual swin transformer u-net for medical image segmentation. IEEE Transactions on Instrumentation and Measurement. 2022;71:1–15. [Google Scholar]
- Lin et al. (2017).Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S. Feature pyramid networks for object detection. Proceedings of the IEEE conference on computer vision and pattern recognition; Piscataway. 2017. [Google Scholar]
- Liu, Hu & Chen (2023).Liu X, Hu Y, Chen J. Hybrid CNN-Transformer model for medical image segmentation with pyramid convolution and multi-layer perceptron. Biomedical Signal Processing and Control. 2023;86:105331. doi: 10.1016/j.bspc.2023.105331. [DOI] [Google Scholar]
- Liu et al. (2021).Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B. Swin transformer: hierarchical vision transformer using shifted windows. Proceedings of the IEEE/CVF international conference on computer vision; Piscataway. 2021. pp. 10012–10022. [Google Scholar]
- Long, Shelhamer & Darrell (2015).Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition; Piscataway. 2015. [DOI] [PubMed] [Google Scholar]
- Lv et al. (2020).Lv Y, Ma H, Li J, Liu S. Attention guided U-Net with atrous convolution for accurate retinal vessels segmentation. IEEE Access. 2020;8:32826–32839. doi: 10.1109/ACCESS.2020.2974027. [DOI] [Google Scholar]
- Ma, Gu & Wang (2024).Ma C, Gu Y, Wang Z. TriConvUNeXt: a pure CNN-Based lightweight symmetrical network for biomedical image segmentation. Journal of Imaging Informatics in Medicine. 2024 doi: 10.1007/s10278-024-01116-8. Epub ahead of print Apr 23 2024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Merskey (1986).Merskey HE. Classification of chronic pain: descriptions of chronic pain syndromes and definitions of pain terms. Pain. 1986;3:S1–S226. [PubMed] [Google Scholar]
- Milletari, Navab & Ahmadi (2016).Milletari F, Navab N, Ahmadi S-A. V-net: fully convolutional neural networks for volumetric medical image segmentation. 2016 fourth international conference on 3D vision (3DV); Piscataway. 2016a. pp. 565–571. [Google Scholar]
- Oktay et al. (2018).Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla NY, Kainz B. Attention u-net: learning where to look for the pancreas. 20181804.03999
- Pacik, Nelson & Werner (2008).Pacik PT, Nelson CE, Werner C. Pain control in augmentation mammaplasty using indwelling catheters in 687 consecutive patients: data analysis. Aesthetic Surgery Journal. 2008;28(6):631–641. doi: 10.1016/j.asj.2008.09.001. [DOI] [PubMed] [Google Scholar]
- Pan et al. (2022).Pan X, Ge C, Lu R, Song S, Chen G, Huang Z, Huang G. On the integration of self-attention and convolution. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; Piscataway. 2022. pp. 815–825. [Google Scholar]
- Pan et al. (2019).Pan X, Li L, Yang D, He Y, Liu Z, Yang H. An accurate nuclei segmentation algorithm in pathological image based on deep semantic network. IEEE Access. 2019;7:110674–110686. doi: 10.1109/ACCESS.2019.2934486. [DOI] [Google Scholar]
- Parmar et al. (2018).Parmar N, Vaswani A, Uszkoreit J, Łukasz Kaiser, Shazeer N, Ku A, Tran D. Image transformer. ArXiv. 2018. [DOI]
- Pham-Dang et al. (2003).Pham-Dang C, Kick O, Collet T, Gouin F, Pinaud M. Continuous peripheral nerve blocks with stimulating catheters. Regional Anesthesia and Pain Medicine. 2003;28(2):83–88. doi: 10.1097/00115550-200303000-00002. [DOI] [PubMed] [Google Scholar]
- Qin et al. (2022).Qin Z, Sun W, Deng H, Li D, Wei Y, Lv B, Yan J, Kong L, Zhong Y. cosFormer: rethinking softmax in attention. 20222202.08791
- Qin et al. (2020).Qin X, Zhang Z, Huang C, Dehghan M, Zaiane OR, Jagersand M. U2-Net: going deeper with nested U-structure for salient object detection. Pattern Recognition. 2020;106:107404. doi: 10.1016/j.patcog.2020.107404. [DOI] [Google Scholar]
- Reza (2004).Reza AM. Realization of the contrast limited adaptive histogram equalization (CLAHE) for real-time image enhancement. Journal of VLSI Signal Processing Systems for Signal, Image and Video Technology. 2004;38:35–44. doi: 10.1023/B:VLSI.0000028532.53893.82. [DOI] [Google Scholar]
- Ronneberger, Fischer & Brox (2015).Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells W, Frangi A, editors. Medical image computing and computer-assisted intervention—MICCAI 2015. MICCAI 2015. Lecture Notes in Computer Science, vol 9351; Cham. 2015. pp. 234–241. [DOI] [Google Scholar]
- Schnabel et al. (2013).Schnabel A, Meyer-Frießem C, Zahn P, Pogatzki-Zahn E. Ultrasound compared with nerve stimulation guidance for peripheral nerve catheter placement: a meta-analysis of randomized controlled trials. British Journal of Anaesthesia. 2013;111(4):564–572. doi: 10.1093/bja/aet196. [DOI] [PubMed] [Google Scholar]
- Sola et al. (2012).Sola C, Raux O, Savath L, Macq C, Capdevila X, Dadure C. Ultrasound guidance characteristics and efficiency of suprazygomatic maxillary nerve blocks in infants: a descriptive prospective study. Pediatric Anesthesia. 2012;22(9):841–846. doi: 10.1111/j.1460-9592.2012.03861.x. [DOI] [PubMed] [Google Scholar]
- Vaswani et al. (2017).Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. 31st conference on neural information processing systems (NIPS 2017), Long Beach, CA, USA.2017. [Google Scholar]
- Vaswani et al. (2023).Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I. Attention is all you need. 20231706.03762
- Wang & Ma (2023).Wang Z, Ma C. Dual-contrastive dual-consistency dual-transformer: a semi-supervised approach to medical image segmentation. Proceedings of the IEEE/CVF international conference on computer vision; Piscataway. 2023. pp. 870–879. [Google Scholar]
- Wang, Shen & Zhou (2019).Wang R, Shen H, Zhou M. Ultrasound nerve segmentation of brachial plexus based on optimized ResU-Net. 2019 IEEE international conference on imaging systems and techniques (IST); Piscataway. 2019. pp. 1–6. [Google Scholar]
- Wang & Voiculescu (2021).Wang Z, Voiculescu I. Quadruple augmented pyramid network for multi-class COVID-19 segmentation via CT. 2021 43rd annual international conference of the IEEE engineering in medicine & biology society (EMBC); Piscataway. 2021. pp. 2956–2959. [DOI] [PubMed] [Google Scholar]
- Wang, Zhang & Voiculescu (2021).Wang Z, Zhang Z, Voiculescu I. RAR-U-Net: a residual encoder to attention decoder by residual connections framework for spine segmentation under noisy labels. 2021 IEEE international conference on image processing (ICIP); Piscataway. 2021. pp. 21–25. [Google Scholar]
- Wang, Zhao & Ni (2022).Wang Z, Zhao W, Ni Z. Adversarial vision transformer for medical image semantic segmentation with limited annotations. BMVC; 2022. p. 1002. [Google Scholar]
- Wijayasinghe et al. (2016).Wijayasinghe N, Duriaud HM, Kehlet H, Anderson KG. Ultrasound guided intercostobrachial nerve blockade in patients with persistent pain after breast cancer surgery: a pilot study. Pain Physician. 2016;19(2):e309. doi: 10.36076/ppj/2016.19.E309. [DOI] [PubMed] [Google Scholar]
- Xie et al. (2021).Xie E, Wang W, Yu Z, Anandkumar A, Alvarez JM, Luo P. SegFormer: simple and efficient design for semantic segmentation with transformers. 35th conference on neural information processing systems (NeurIPS 2021).2021. [Google Scholar]
- Yang & Yang (2023).Yang H, Yang D. CSwin-PNet: a CNN-Swin transformer combined pyramid network for breast lesion segmentation in ultrasound images. Expert Systems with Applications. 2023;213:119024. doi: 10.1016/j.eswa.2022.119024. [DOI] [Google Scholar]
- Yang et al. (2016).Yang Z, Yang D, Dyer C, He X, Smola A, Hovy E. Hierarchical attention networks for document classification. Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies; 2016. pp. 1480–1489. [Google Scholar]
- Yu, Koltun & Funkhouser (2017).Yu F, Koltun V, Funkhouser T. Dilated residual networks. Proceedings of the IEEE conference on computer vision and pattern recognition; Piscataway. 2017. [Google Scholar]
- Zhang et al. (2024).Zhang H, Lian J, Yi Z, Wu R, Lu X, Ma P, Ma Y. HAU-Net: hybrid CNN-transformer for breast ultrasound image segmentation. Biomedical Signal Processing and Control. 2024;87:105427. doi: 10.1016/j.bspc.2023.105427. [DOI] [Google Scholar]
- Zhang et al. (2023).Zhang L, Lu J, Zhang J, Zhu X, Feng J, Xiang T. Softmax-free linear transformers. 20232207.03341v3
- Zhou, He & Jia (2020).Zhou Z, He Z, Jia Y. AFPNet: a 3D fully convolutional neural network with atrous-convolution feature pyramid for brain tumor segmentation via MRI images. Neurocomputing. 2020;402:235–244. doi: 10.1016/j.neucom.2020.03.097. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The following information was supplied regarding data availability:
The code is available in the Supplemental File.
The raw data is publicly available at Kaggle: https://www.kaggle.com/competitions/ultrasound-nerve-segmentation.