Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2022 Sep 14;2022:4259471. doi: 10.1155/2022/4259471

Automatic Segmentation of Lumbar Spine MRI Images Based on Improved Attention U-Net

Shuai Wang 1, Zhengwei Jiang 1, Hualin Yang 1, Xiangrong Li 1, Zhicheng Yang 2,
PMCID: PMC9492365  PMID: 36156962

Abstract

Lumbar spine segmentation is important to help doctors diagnose lumbar disc herniation (LDH) and patients' rehabilitation treatment. In order to accurately segment the lumbar spine, a lumbar spine image segmentation algorithm based on improved Attention U-Net is proposed. The algorithm is based on Attention U-Net, the attention module based on multilevel feature map fusion is adopted, two residual modules are introduced instead of the original convolution blocks. a hybrid loss function is used for prediction during the training process, and finally, the image superposition process is realized. In this experiment, we expanded 420 lumbar MRI images of 180 patients to 1000 images and trained them by different algorithms, respectively, and accuracy, recall, and Dice similarity coefficient metrics were used to analyze these algorithms. The results show that compared with SVM, FCN, R-CNN, U-Net, and Attention U-Net models, the improved model achieved better results in all three evaluations, with 95.50%, 94.53%, and 95.01%, respectively, which proves the better performance of the proposed method for segmentation in lumbar disc and caudal vertebrae.

1. Introduction

In recent years, lumbar spine-related diseases have been affecting people's normal work and lives, and some families are bearing a huge economic burden. Lumbar spine diseases mainly include disc herniation (LDH) and lumbar spinal stenosis [1]. As LDH is mainly caused by disc degeneration or overwork, it has the highest prevalence in the age group of 30–50 years old, with a prevalence ratio of 2 : 1.1 between men and women, and 90% of the elderly over 60 years old worldwide suffer from degenerative disc symptoms [25]. Due to the increased pressure in life, more and more young people are also suffering from lumbar spine diseases. Lumbar spine diseases are diagnosed by physicians by examining the relevant parts of the lumbar spine. The imaging modalities mainly include computed tomography (CT) [6], magnetic resonance imaging (MRI) [7], and so on. The images with better imaging are selected from lots of CT or MRI images by physicians, and the possible lesions are diagnosed by physicians. As the number of patients increases, the cumbersome diagnostic approach not only increases the consultation time of patients but also puts tremendous work pressure on physicians. Therefore, lumbar segmentation plays an important part in the whole diagnostic process, which helps doctors to observe medical images quickly and accurately, and it facilitates the patient's further treatment.

Medical image segmentation techniques can be divided into two categories: traditional segmentation techniques and deep learning-based methods. The former includes threshold-based segmentation [8], edge-based segmentation [9], region-based segmentation [10, 11], and active contour model-based techniques [12, 13], and the latter is mainly neural network-based segmentation [1418]. Earlier studies have been applied to the lumbar spine segmentation by traditional segmentation techniques. Hoad et al. [19] used a traditional threshold segmentation method applied to the spine MRI images to segment lumbar discs from soft tissues, thereby realizing the computer-aided diagnosis of the spine. Armato et al. [20] demonstrated a random forest based method to extract vertebral height and width, which solved the problem of biomarkers of the lumbar spine. Punarselvam et al. [21] used a watershed method to detect boundaries and edges in the lumbar spine images. Later, this method was added with statistical and spectral texture features by some scholars [22], and they used this method to effectively distinguish the closed area of the intervertebral disc in the image. In recent years, deep learning has shown great advantages in medical image processing. Lee et al. [23] established a reference framework for segmenting lumbar arch roots in CT images, which obtained segmented vertebrae and canal references by using 2D dynamic thresholding and a combined cost based on sparing and finally achieved edge segmentation of the spine. Deng et al. [24] proposed a method based on the combination of contour transform and artificial neural network (ANN). This approach used contour transform to decompose images to obtain contour coefficients and used the ANN to optimize the coefficients of contour transform, thereby improving the performance of lumbar image segmentation. In addition, the method of the deep network with full convolution (FCN) was proposed in some studies [25, 26], which segmented and labeled the lumbar spine at once by using the local lumbar spine environment. FCN combined with the convolutional neural network (CNN) again to improve the segmentation effect, compared the segmentation results with conventional segmentation methods, and the results showed that the segmentation accuracy and efficiency were improved [27]. However, deep convolutional networks are not the only option for medical image segmentation, which requires training with a large amount of data. With the expansion of machine learning, image segmentation has developed different types of models based on full convolutional networks, such as U-Net [2830], PSPNet [31, 32], and DeepLab [33, 34]. Sunetra et al. [35] showed the LDS U-Net structure to segment ultrasound spine lateral bony features from noisy images, which required only a small number of medical images for training. Saenz-Gamboa et al. [36] used a variant of U-Net for automatic segmentation of lumbar spine MRI images; this model classified labels to each pixel of the image. For the image noise problem of lumbar spine segmentation, Yang et al. [37] showed an automatic initialization level set method based on regional correlation, which introduced the histogram information inside and outside the level set contour, and Tang et al. [38] used a double densely connected U-neural network. This method improved the contrast of vertebral body edges, spinal ducts, and cloudy sacs while reducing image noise.

The lumbar spine can be extracted from the soft tissue by the above lumbar spine segmentation methods, but the accuracy is still slightly low, and the segmentation effect is influenced by the lesion area, as well as the parameter settings, which has some limitations. To solve these problems, this paper proposes a lumbar spine segmentation method with an improved Attention U-Net, which improves the structure of the attention module and residual network. A hybrid loss function is used to improve the detection accuracy. The specific experimental procedure is shown in Figure 1. First, the images are preprocessed by extracting local binary pattern (LBP) features and using contrast limited adaptive histogram equalization (CLAHE). Then, segmentation is performed using a modified Attention U-Net model. Finally, the vertebral blocks and intervertebral discs are extracted by gray threshold, and the images are superimposed by image fusion. MRI is the mainstay of lumbar spine image diagnosis at present. Compared with CT, MRI images have clearer soft tissue contours and have a better effect for imaging intervertebral disc degeneration, so we select MRI images to carry out the experiment of lumbar spine segmentation.

Figure 1.

Figure 1

Segmentation process of lumbar spine image.

The rest of this article is organized as follows: in Section 2, U-Net and Attention U-Net are introduced, and how to improve Attention U-Net is described in detail. The arrangement of the experiments and the evaluation metrics are described in Section 3. The results of the experiments and postprocessing of the images are described in Section 4. Finally, Section 5 draws conclusions and proposes future research.

2. Methods

2.1. U-Net and Attention U-Net

U-Net is a convolutional neural network architecture with a simple structure and high efficiency. The architecture consists of two parts: an encoder and a decoder. The encoder part uses convolution and pooling to downsample the image, which doubles the number of feature channels and halves the image size. This part consists of two convolutional layers with 3×3 filters and a 2×2 maximum pooling layer with a step size of 2. ReLU is used for the activation function. The decoder part upsamples the feature image by deconvolution, which reduces the number of features channels and increases the image size, and finally outputs an image of the same size as the original image. It mainly consists of a deconvolutional layer with 2×2 filter and two 3×3 convolutional layers and still uses the ReLU activation function. The feature stitching of U-Net has a better processing effect for problems such as the difficulty in distinguishing biological tissue structures and the display of low-level and high-level features [39]. In addition, the experimental data of some medical images are generally so less that they are not suitable for the complex and large networks, while the U-Net with its simple structure can be better processed for these medical images.

Attention U-Net is a network structure based on U-Net with an added attention mechanism [40]. Compared with U-Net, an attention mechanism is added to the feature map in the encoder part before splicing in the decoder part, so that irrelevant background regions are suppressed and target regions are enhanced. In the lumbar segmentation, the vertebral body, intervertebral disk, and sacral regions are enhanced by the attention mechanism, while the soft tissue regions are suppressed [41]. The oversegmentation of images by the network structure can be effectively reduced by the attention mechanism.

2.2. Improved Attention U-Net

In this study, an improved network structure is proposed based on Attention U-Net, as shown in Figure 2. The network is presented as U-shaped, with the encoder part on the left and the decoder part on the right side. The number of channels, height, and width of the tensor are denoted by D, H, and W, respectively. Compared with the traditional Attention U-Net, the network framework is built by deep convolution in the bottom feature layer and the top feature layer. Improved residual structure is added to the convolution process of each layer to increase the depth and feature fusion ability of the network. Layers 1–3 in the encoding process and layers 6–8 in the decoding process are connected by jump connections, so that the encoder part is used to generate feature information at different scales in the whole network. An improved attention module is introduced for each jump connection that allows the model to acquire local information more accurately. Finally, the structure reduces the number of four downsampling layers in the traditional U-Net to three, which reduces the number of parameters in the network; therefore, the computational complexity is reduced, which facilitates the acquisition of global features.

Figure 2.

Figure 2

The network structure of improved Attention U-Net.

2.2.1. Improved Multilevel Attention Module

The attention mechanism is generally applied in the dynamic analysis of vision and classification of images and later in segmentation of images. In image segmentation, the attention mechanism is used to remove redundant information from layers to improve the running speed and segmentation performance of matrix algorithms [42, 43]. The expression of this attention mechanism is

AGx,g,g=σConv2d1×1ReLuConv2d1×1Upg×Conv2d1×1x×x, (1)

where the eigengraphs of the encoder output and the gating signal are represented as x and g, respectively, σ is the sigmoid function, Up represents the upsampling, and Conv2 d1×1represents the two-dimensional 1 × 1 convolution.

According to the characteristics of attention, an attention module based on a multilevel feature map fusion is improved in this study. The improved attention module is shown in Figure 3; in the input phase, the matrix of the encoder part is normalized by conv1×1 and batch normalization (BN) operation in the input stage, combined with the matrix in the upsampling that has undergone convolution and batch normalization, and then, the convergence of the attention parameters is trained by processing the ReLU activation function, the conv1 × 1, and sigmoid activation functions. The attention coefficients α (i.e., attention weights) are obtained by resampling. Finally, the output α is multiplied with the feature layer of the encoder to obtain the result.

Figure 3.

Figure 3

Improved attention module.

2.2.2. Improved Residual Module

Different from the conventional convolutional neural network, the ResNet [44] residual network deepens the number of network layers through shortcut connections. It still has better running speed and results without adding parameters and data calculations, which can effectively solve the gradient dissipation problem caused by too many output features. The residual network is composed of residual modules, which are as follows:

y=Fx,wi+wjx, (2)

where y and x are the output and input vectors of the residual module, F(x, {wi}) is the residual mapping, and linear projection wj is used for matching dimensions in shortcut connection.

This study designs two different residual modules based on the ResNet network structure to replace the convolutional blocks in Attention U-Net, and the module structures are shown in Figures 3 and4.

Figure 4.

Figure 4

Improved standard convolutional residual blocks.

The convolutional module in Figure 4 is used for feature extraction in the first and last layers of Attention U-Net, adding the underlying residual structure to better extract information in large size and shallow-depth feature maps. Its two convolutional layers use a 3 × 3 convolutional kernel and ReLU function, and the input feature map after 1 × 1 convolution is subjected to a feature summation operation with the output after 3 × 3 convolution. In addition, a batch normalization (BN) operation is performed before using ReLU to speed up the convergence of the model.

An improved deep convolutional residual module is shown in Figure 5. It mainly contains two 5 × 5 convolutional layers, a 3 × 3 convolution, and some basic operations. For the feature map that is input to the residual module, the feature map after two 5 × 5 deep convolutional operations is fused to form a new feature map and stitched, at which time the number of channels becomes twice as many as the original one, fusing feature information of different complexity. Then, it is fused with the output feature map after 3 × 3 convolution, batch normalization, and ReLU operation, and finally, the feature map is input to the next residual model. The residual module introduced in this module enhances the feature extraction capability at different depths; two 5 × 5 depth convolutional layers are able to extract semantic information at different levels of complexity, which are adopted to the feature extraction stage of the high-level feature map. So, this module is used instead of the convolutional layer in the original Attention U-Net network.

Figure 5.

Figure 5

Improved deep convolutional residual blocks.

2.2.3. Hybrid Loss Function

The loss function optimizes the network structure by backpropagating the numerical error of the calculated loss function and continuously updating the weights. In the field of medical image segmentation, the Dice Loss [45] function is commonly used to calculate the degree of differences between the predicted region and the real region. The concepts of Dice Loss are defined by (3). However, the loss function has the problem that the training error curve is very confusing when using the Dice Loss or IOU [46] loss function causes. These situations can be avoided by using the cross-entropy loss (CEL) function, with the expression (4), which makes the gradient form better, but suffers from the problem of class imbalance.

D=12XY/X+Y, (3)
CL=i=1nyicilogyic^i, (4)

where XY denotes the intersection of X and Y; |X| and |Y| denote the number of X and Y, respectively; and yiciand yic^idenote the label value and the predicted value, respectively.

To address the above problems, we use a hybrid loss function based on cross-entropy loss function and Dice loss function, which observes the convergence steadily during the training process and avoids the category of the imbalanced problem. Its formula is as follows:

DCLQ,F=1UH=1HU=1Uyn,llogpnl+2yn,lpnlyn,l2pn,l2, (5)

where Q is the actual situation, F is the predicted result from training, yn,lF represents the probability of prediction, the pnlQ represents the established target, H represents the number of classifications in the dataset, and U represents the number of pixels in the image.

3. Experiment

3.1. Experimental Data and Environment

The experimental dataset used in this study is collected from Qingdao Hospital of Shandong University Qilu Hospital. The collected data contain 420 MRI T1 images of 180 lumbar spine patients, which were expanded to 1000 images by data enhancement. In order to facilitate the later experimental operation, the dcm format of T1 images is changed to jpg format by the SimpleITK toolkit, and the image size is resized to 512 × 512 by linear interpolation. Finally, the images are classified into training sets, validation sets, and test sets according to the ratio of 7 : 2:1.

Experimental environment: Windows 10 operating system, AMD Ryzen 7 4800H processor, NVIDIA GeForce GTX 1650 GPU with 64 GB of video memory, 8-core CPU, 32 G of memory, PyTorch 1.2.0 is used for the deep learning framework, and Python is used for the programming language.

3.2. Data Preprocessing

The MRI images of the lumbar spine contain tissues such as vertebrae, intervertebral discs, spinal canal, and muscles.

The texture features of tissues reflect different basic feature information. Because the intervertebral disc in the original image is relatively dark and the contrast with the vertebral block is not obvious, it is easy to cause errors in the manual labeling of the intervertebral disc and vertebral block in the later stages, making the labeled image inaccurate and affecting the effect of later deep learning. Therefore, in order to facilitate labeling, the contrast of different tissues in the image needs to be improved. The histogram equalization method can increase the gray value range of the image and uniformize the distribution of pixel gray values, thereby improving the contrast and clarity of the image. The specific steps are as follows: first, the grayscale values are calculated, and the histograms are counted. Then, the cumulative histogram in the statistical histogram is calculated, and finally, the interval conversion is performed on the cumulative histogram. The effect is shown in Figure 6(b). The figure shows that after the histogram transformation of the whole image by the conventional histogram equalization operation, the brightness of the image is improved, the image noise is amplified, and even some parts appear. The effect of excessive brightness is that it eventually leads to a decrease in the sharpness of the image.

Figure 6.

Figure 6

The comparison results of equalization. (a) Original graph. (b) Histogram equalization. (c) Our method.

In order to effectively solve the problem of the image noise signal being amplified, a method of limiting contrast is added to the adaptive histogram equalization: if the value in a certain range of the histogram exceeds the limited threshold, the exceeding area will be cut out and that area is distributed to the rest of the histogram. As shown in Figure 6, compared with conventional histogram equalization, the method used in this study has a better processing effect and improves the contrast and sharpness of different tissues.

The processed images are labeled and classified by the LabelMe tool; each image is labeled with 3 categories, including 5 vertebral bodies, 5 intervertebral discs, and 1 sacrum, named L, LD, and S. The different tissues labeled are distinguished by different colors, as shown in Figure 7. After the image is annotated, the corresponding JSON format file is obtained, which contains information about the different categories of annotation and the pixel coordinates of the annotation points. The JSON file is transformed into voc data, which contain the segmented mask map, the annotation map combining the mask map with the original map, and the NPY format file.

Figure 7.

Figure 7

Marked effect.

3.3. Evaluation Indicators

To comprehensively evaluate the effectiveness of the proposed algorithm for lumbar spine segmentation, three metrics are used as evaluation methods, which are accuracy (P), recall (R), and Dice similarity coefficient. As shown in formulas (6) and (7), when the accuracy, P, is high, the recall, R, is low and vice versa. When the two do not conform to this relationship, the Dice similarity coefficient is introduced for comprehensive evaluation. The Dice similarity coefficient is often used to evaluate medical image targets with uneven segmentation size. The obtained formula is (8).

P=TPTP+FP, (6)
R=TPTP+FN, (7)
Dice=2TPTP+FP+TN+FN, (8)

where true positive (TP) is the number of lumbar spine images that the model correctly classifies as positive examples, that is, the number of samples that are actually positive examples and are classified as positive examples by the model. False positives (FP), which indicates the number of lumbar spine images that the model incorrectly classifies as positive, that is, the number of samples that are actually negative but are classified as positive by the model. True negative (TN) is the number of lumbar spine images correctly classified as negative by the model, that is, the number of samples that are actually negative and are classified as negative by the model. False negative (FN), which indicates the number of lumbar spine images that the model incorrectly classifies as negative examples, that is, the number of samples that are actually positive examples but are classified as negative examples by the model.

4. Results and Analysis

4.1. Attention U-Net Segmentation Effect

During the training of the algorithm model, the batch size is set to 2, and 15 rounds are trained with one validation per round as well as model preservation. In this paper, the models of the Attention U-Net network with improved residual module, attention mechanism, and hybrid loss function are defined as R-Attention U-Net, A-Attention U-Net, and L-Attention U-Net, respectively. As shown in Figure 8, with the continuous iteration of R-Attention U-Net, A-Attention U-Net, L-Attention U-Net, and the improved model in this study, the loss function gradually decreases. At the beginning of training, after three rounds, the loss function of all models rapidly decreases to below 0.4. After the end of training, the loss functions of all models were stabilized below 0.3, and segmentation models satisfying the requirements were obtained. Table 1 shows the experimental results of different models for lumbar spine detection. The accuracy rates of R-Attention U-Net, A-Attention U-Net, and L-Attention U-Net models are all above 90%, which indicate that the methods for improving a single variable all have better performance. Among them, A-Attention U-Net has the highest recall rate of 95.70%, but the number of incorrectly identified samples is high, resulting in a low accuracy rate. The improved Attention U-Net in this study has a lower recall rate compared with the A-Attention U-Net with the improved attention mechanism only, but the accuracy and Dice similarity coefficient are improved by 2.23% and 0.54%, respectively, i.e., the incorporated residual module and hybrid loss function have a better correction effect for the incorrect identification of the A-Attention U-Net.

Figure 8.

Figure 8

The loss curve.

Table 1.

Comparison of experimental results of different improvement schemes.

Type TP FP FN P (%) R (%) Dice (%)
R-Attention U-Net 728 46 47 94.06 93.94 93.99
A-Attention U-Net 735 53 33 93.27 95.70 94.47
L-Attention U-Net 722 49 50 93.64 93.52 93.58
Improved Attention U-Net 743 35 43 95.50 94.53 95.01

Table 2 shows the comparison of experimental results under equal conditions; SVM, FCN, R–CNN, U-Net, Attention U-Net, and improved Attention U-Net are used for samples segmentation, respectively; the improved model in this study achieves the best results in terms of accuracy, recall, and Dice similarity coefficient indexes, which are 95.50%, 94.53%, and 95.01%, proving that the proposed method meets the requirements of lumbar spine segmentation. Among them, R-CNN and Attention U-Net also have better recognition effects, and by comparing the experimental results of these two models and the method in this study, as shown in Figure 9, the segmentation effect of the algorithm in this study is significantly better than other methods in terms of overall and detail processing. The sacrum and intervertebral disc appear missing and broken by R-CNN and Attention U-Net processing, while the proposed algorithm can segment the feature edges of the target more accurately, reducing the occurrence of problems, such as the mutilation of the intervertebral disc and the missing sacrum, and showing better robustness to targets with poor clarity and different shapes.

Table 2.

Comparison of experimental results of different algorithms.

Type TP FP FN P (%) R (%) Dice (%)
SVM 605 101 115 85.69 84.03 84.85
FCN 644 83 94 88.58 87.27 87.91
R-CNN 712 58 51 92.50 93.32 92.89
U-Net 685 62 74 91.70 90.25 90.10
Attention U-Net 715 49 57 93.59 92.62 93.10
Improved Attention U-Net 743 35 43 95.50 94.53 95.01

Figure 9.

Figure 9

The segmentation effect of different algorithms.

4.2. Postprocessing of Segmented Images

Postprocessing of the images is required to observe the different parts of the lumbar spine more clearly. As shown in Figure 10, different categories are represented by different colors, with the vertebral block represented by green, the intervertebral disc represented by yellow, the sacrum represented by blue, and the background represented by red. The extraction of each category is performed according to the different colors of the segmentation object. First, the mask image is converted to a grayscale image, and then, the grayscale value threshold is set by the different grayscale value sizes of the vertebral block and intervertebral disc, so that the vertebral block and intervertebral disc can be extracted. Figure 10 shows the effect of extracting the vertebral block alone.

Figure 10.

Figure 10

The segmentation effect of different algorithms.

In order to realize the detailed view of different parts, we achieve the superposition of the original lumbar spine image and the mask image by image fusion and add labels to the hybrid map. Figure 10 shows the fusion map of the segmented images and the original image and the fusion map of the edge extraction and the original image, respectively. We can clearly see the various parts of the lumbar spine and lumbar disc from the fusion map, which is beneficial for the doctor to quickly diagnose the lumbar spine MRI images.

5. Conclusions and Future Work

Lumbar spine segmentation is very important for the diagnosis of related diseases. To address the problem of low segmentation accuracy of lumbar spine MRI images, we propose a segmentation method based on improved Attention U-Net. The steps of the study are as follows:

  1. Limiting contrast is added to the adaptive histogram equalization, which reduces the roughness of the image and improves the contrast and sharpness of different tissues, thus facilitating the labeling of experimental data

  2. By improving Attention U-Net, two residual modules are introduced instead of the original convolutional blocks, an attention module based on a multilevel feature map fusion is used, and a hybrid loss function is used in training for prediction

  3. Different tissues are extracted according to the different colors of the segmented images. And through image fusion, the superposition of the original lumbar spine image and the segmented image is realized, thus facilitating the physician to observe the lesion of each tissue more intuitively.

According to the comparison experiments of the three models with changing single variable, among them, the recall rate of A-Attention U-Net performs better than the improved method in this paper, reaching 95.70%, but the false recognition rate of A-Attention U-Net is higher, which leads to a decrease in accuracy and Dice similarity coefficient by 2.23% and 0.54%, respectively, proving that the method in this study is better than the improved single-variable method with better equalization ability. In addition, comparison experiments of six different network models were completed, and it was verified that the model in this study has better results in lumbar spine segmentation and outperforms SVM, FCN, R-CNN, U-Net, and Attention U-Net in terms of accuracy, recall, and Dice similarity coefficient, with 95.50%, 94.53%, and 95.01%, respectively. It proves that the method in this study has better performance in the intervertebral disc and more detailed processing of the sacral region with better robustness.

Acknowledgments

This research was funded by the National Natural Science, China (52101401).

Data Availability

The datasets used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  • 1.Hoy D., March L., Brooks P., et al. The global burden of low back pain: estimates from the Global Burden of Disease 2010 study. Annals of the Rheumatic Diseases . 2014;73(6):968–974. doi: 10.1136/annrheumdis-2013-204428. [DOI] [PubMed] [Google Scholar]
  • 2.Hoy D., March L., Brooks P., et al. Measuring the global burden of low back pain. Best Practice & Research Clinical Rheumatology . 2010;24(2):155–165. doi: 10.1016/j.berh.2009.11.002. [DOI] [PubMed] [Google Scholar]
  • 3.Ponnappan R. K., Markova D. Z., Antonio P. J., et al. An organ culture system to model early degenerative changes of the intervertebral disc. Arthritis Research and Therapy . 2011;13(5):1711–R212. doi: 10.1186/ar3494. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Pfirrmann C. W. A., Metzdorf A., Zanetti M., Hodler J., Boos N. Spine . 2001;26(17):1873–1878. doi: 10.1097/00007632-200109010-00011. [DOI] [PubMed] [Google Scholar]
  • 5.Yuan T., Zhang J., Liu Q. Treatment of calcified L5S1 lumbar disc herniation with percutaneous endoscopic interlaminar discectomy:A report of 15 cases and literature review. Journal of Jilin University - Medicine Edition . 2018;44:615–619. [Google Scholar]
  • 6.Bae H. J., Hyun H., Byeon Y., et al. Fully automated 3D segmentation and separation of multiple cervical vertebrae in CT images using a 2D convolutional neural network. Computer Methods and Programs in Biomedicine . 2020;184 doi: 10.1016/j.cmpb.2019.105119.105119 [DOI] [PubMed] [Google Scholar]
  • 7.Hoad C. L., Martel A. L. Segmentation of MR images for computer-assisted surgery of the lumbar spine. Physics in Medicine and Biology . 2002;47(19):3503–3517. doi: 10.1088/0031-9155/47/19/305. [DOI] [PubMed] [Google Scholar]
  • 8.Angulakshmi M., Deepa M. A review on deep learning architecture and methods for MRI Brain Tumour segmentation. Current Medical Imaging Formerly Current Medical Imaging Reviews . 2021;17:695–706. doi: 10.2174/1573405616666210108122048. [DOI] [PubMed] [Google Scholar]
  • 9.Park J., Park S., Cho W. Medical image segmentation using level set method with a new hybrid speed function based on boundary and region segmentation. IEICE Transactions on Information and Systems . 2012;E95.D(8):2133–2141. doi: 10.1587/transinf.e95.d.2133. [DOI] [Google Scholar]
  • 10.Geweid G. G. N., Abdallah M. A. A novel approach for Breast Cancer Investigation and recognition using M-level set-based Optimization functions. IEEE Access . 2019;7:136343–136357. doi: 10.1109/access.2019.2941990. [DOI] [Google Scholar]
  • 11.Huang Y., Hu G., Ji C., Xiong H. Glass-cutting medical images via a mechanical image segmentation method based on crack propagation. Nature Communications . 2020;11(1):p. 5669. doi: 10.1038/s41467-020-19392-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Afifi A., Nakaguchi T., Tsumura N., Miyake Y. A model Optimization approach to the automatic segmentation of medical images. IEICE Transactions on Information and Systems . 2010;E93-D(4):882–890. doi: 10.1587/transinf.e93.d.882. [DOI] [Google Scholar]
  • 13.Zhang L., Liu J., Shang F. X., Li G., Zhao J., Zhang Y. Robust segmentation method for noisy images based on an unsupervised denosing filter. Tsinghua Science and Technology . 2021;26(5):736–748. doi: 10.26599/tst.2021.9010021. [DOI] [Google Scholar]
  • 14.Ma Y., Li X., Duan X., Peng Y., Zhang Y. Retinal Vessel segmentation by deep residual learning with Wide activation. Computational Intelligence and Neuroscience . 2020;2020:11. doi: 10.1155/2020/8822407.8822407 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Hesamian M. H., Jia W., He X., Kennedy P. Deep learning techniques for medical image segmentation: Achievements and Challenges. Journal of Digital Imaging . 2019;32(4):582–596. doi: 10.1007/s10278-019-00227-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Chen C., Qin C., Qiu H., et al. Deep learning for Cardiac image segmentation: a review. Frontiers in Cardiovascular Medicine . 2020;7:25–36. doi: 10.3389/fcvm.2020.00025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Seo H., Badiei Khuzani M., Vasudevan V., et al. Machine learning techniques for biomedical image segmentation: an Overview of Technical Aspects and introduction to State-of-Art Applications. Medical Physics . 2020;47(5):148–167. doi: 10.1002/mp.13649. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Asgari Taghanaki S., Abhishek K., Cohen J. P., Cohen-Adad J., Hamarneh G. Deep semantic segmentation of natural and medical images: a review. Artificial Intelligence Review . 2020;54(1):137–178. doi: 10.1007/s10462-020-09854-1. [DOI] [Google Scholar]
  • 19.Eckstein F., Cicuttini F., Raynauld J. P., Waterton J. C., Peterfy C. Magnetic resonance imaging (MRI) of articular cartilage in knee osteoarthritis (OA): Morphological assessment. Osteoarthritis and Cartilage . 2006;14:46–75. doi: 10.1016/j.joca.2006.02.026. [DOI] [PubMed] [Google Scholar]
  • 20.Armato S. G., Petrick N. A., Kulkarni A., et al. Automatic Segmentation of Lumbar Vertebrae in CT Images . medical Imaging; 2017. [Google Scholar]
  • 21.Punarselvam E., Suresh D. P., Parthasarathy R., Suresh M. Segmentation of lumbar spine image using watershed algorithm. Journal of Engineering Research and Applications . 2013;3:1386–1389. [Google Scholar]
  • 22.Chevrefils C., Cheriet F., Aubin C.-E., Grimard G. Texture analysis for automatic segmentation of intervertebral disks of scoliotic spines from MR images. IEEE Transactions on Information Technology in Biomedicine . 2009;13(4):608–620. doi: 10.1109/titb.2009.2018286. [DOI] [PubMed] [Google Scholar]
  • 23.Lee J., Kim S., Kim Y. S., Chung W. K. Automated segmentation of the lumbar Pedicle in CT images for spinal fusion surgery. IEEE Transactions on Biomedical Engineering . 2011;58(7):2051–2063. doi: 10.1109/tbme.2011.2135351. [DOI] [PubMed] [Google Scholar]
  • 24.Deng J. M., Li H. Y., Wu H. An approach to lumbar Vertebra CT image segmentation using contourlet transform and ANNs. Advanced Materials Research . 2012;468-471:613–618. doi: 10.4028/www.scientific.net/amr.468-471.613. [DOI] [Google Scholar]
  • 25.Sekuboyina A., Valentinitsch A., Kirschke J. S., Menze B. H. A Localisation-Segmentation Approach for Multi-Label Annotation of Lumbar Vertebrae Using Deep Nets. 2017. https://arxiv.org/abs/1703.04347.
  • 26.Zeiler M. D., Fergus R. Visualizing and Understanding convolutional networks. Anal. Chem. Res. . 2014;12:818–833. [Google Scholar]
  • 27.Vania M., Mureja D., Lee D. Automatic spine segmentation from CT images using Convolutional Neural Network via redundant generation of class labels. Journal of Computational Design and Engineering . 2019;6(2):224–232. doi: 10.1016/j.jcde.2018.05.002. [DOI] [Google Scholar]
  • 28.Ronneberger O., Fischer P., Brox T. U-net: Convolutional Networks for Biomedical Image Segmentation . Berlin, Germany: Springer International Publishing; 2015. pp. 234–241. [Google Scholar]
  • 29.Ronneberger O., Fischer P., Brox T. U-net: convolutional networks for biomedical image segmentation. coRR . 2015;9351:234–241. [Google Scholar]
  • 30.Awan M. J., Rahim M. S. M., Salim N., Rehman A., Garcia-Zapirain B. Automated knee MR images segmentation of Anterior Cruciate Ligament Tears. Sensors . 2022;22(4):p. 1552. doi: 10.3390/s22041552. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Zhao H., Shi J., Qi X., Wang X., Jia J. Pyramid scene parsing network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); July 2017; Honolulu, HI, USA. pp. 6230–6623. [Google Scholar]
  • 32.Shi B., He H. Q., You Qi. A method of multi-scale Total convolution network Driven Remote Sensing image Repair. Journal of Geomatics . 2018;43:124–126. [Google Scholar]
  • 33.Chen L. C., Papandreou G., Kokkinos I., Murphy K., Yuille A. L. Semantic image segmentation with deep convolutional Nets and fully connected CRFs. Computer Science . 2014;4:357–361. doi: 10.1109/TPAMI.2017.2699184. [DOI] [PubMed] [Google Scholar]
  • 34.Chen L.-C., Papandreou G., Kokkinos I., Murphy K., Yuille A. L. DeepLab: semantic image segmentation with deep convolutional Nets, Atrous convolution, and fully connected CRFs. IEEE Transactions on Pattern Analysis and Machine Intelligence . 2018;40(4):834–848. doi: 10.1109/tpami.2017.2699184. [DOI] [PubMed] [Google Scholar]
  • 35.Banerjee S., Lyu J, Huang Z, et al. Light-convolution Dense selection U-net (LDS U-net) for ultrasound lateral bony feature segmentation. Applied Sciences . 2021;11(21):10180–10198. doi: 10.3390/app112110180. [DOI] [Google Scholar]
  • 36.Saenz-Gamboa J. J., Domenech J., Alonso-Manjarrez A., Gomez J. A., Maria I. V. Automatic semantic segmentation of the lumbar spine. Proceedings of the 25 th International conference Clinical Applicability in a Multi-Parametric and Multi-centre MRI Study; January 2021; Milan, Italy. [Google Scholar]
  • 37.Yang L., Wei L., Tan J., Zhang Y. A novel automatically initialized level set approach based on region correlation for lumbar vertebrae CT image segmentation. Proceedings of the IEEE International Symposium on Medical Measurements & Applications; May 2015; Turin, Italy. IEEE; pp. 291–296. [Google Scholar]
  • 38.Tang H., Pei X., Huang S., Li X., Liu C. Automatic lumbar spinal CT image segmentation with a dual densely connected U-net. IEEE Access . 2020;8:89228–89238. doi: 10.1109/access.2020.2993867. [DOI] [Google Scholar]
  • 39.He K., Zhang X., Ren S., Sun J. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; June 2016; New York, NY, USA. IEEE; pp. 770–778. [Google Scholar]
  • 40.Yin Z., Sun D., Ren T., et al. Research on automatic gallbladder segmentation model based on improved Attention U-Net. Beijing Biomedicine Engineering . 2021;40:346–353+376. [Google Scholar]
  • 41.Gupta A., Upadhyaya S., Yeung C. M., et al. Disk area is a more Reliable Measurement than Anteroposterior Length in the assessment of lumbar disk herniations: a validation study. Clinical Spine Surgery: A Spine Publication . 2020;33(8):381–385. doi: 10.1097/bsd.0000000000000958. [DOI] [PubMed] [Google Scholar]
  • 42.Itti L., Koch C., Niebur E. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence . 1998;20(11):1254–1259. doi: 10.1109/34.730558. [DOI] [Google Scholar]
  • 43.Yin J., Zhou Z., Xu S., Yang R., Liu K. A generative Adversarial network fused with Dual-attention mechanism and its Application in Multitarget image fine segmentation. Computational Intelligence and Neuroscience . 2021;2021:1–16. doi: 10.1155/2021/2464648.2464648 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Liu P., Sun L., Zhang C. Y. Fault Text classification based on Interactive attention mechanism network model. Computer Integrated Manufacturing System . 2021;27:72–89. [Google Scholar]
  • 45.Cai Y., Li Q., Fan Y., Zhang L., Huang H., Ding X. An automatic trough line identification method based on improved UNet. Atmospheric Research . 2021;264 doi: 10.1016/j.atmosres.2021.105839.105839 [DOI] [Google Scholar]
  • 46.Zheng Z., Wang P., Liu W., Li J., Ye R., Ren D. Distance-iou Loss: Faster and Better Learning for Bounding Box Regression. 2020. pp. 12993–13000. https://arxiv.org/abs/1911.08287.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets used to support the findings of this study are available from the corresponding author upon request.


Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES