Abstract
Liver segmentation and recognition from computed tomography (CT) images is a warm topic in image processing which is helpful for doctors and practitioners. Currently, many deep learning methods are used for liver segmentation that takes a long time to train the model which makes this task challenging and limited to larger hardware resources. In this research, we proposed a very lightweight convolutional neural network (CNN) to extract the liver region from CT scan images. The suggested CNN algorithm consists of 3 convolutional and 2 fully connected layers, where softmax is used to discriminate the liver from background. Random Gaussian distribution is used for weight initialization which achieved a distance-preserving-embedding of the information. The proposed network is known as Ga-CNN (Gaussian-weight initialization of CNN). General experiments are performed on three benchmark datasets including MICCAI SLiver'07, 3Dircadb01, and LiTS17. Experimental results show that the proposed method performed well on each benchmark dataset.
1. Introduction
Liver disease is among the most serious medical conditions that can threaten human life and health. Liver tumors are the second primary reason of death in men and the sixth foremost cause of death in females. In 2008, 750,000 people have been identified with liver malignancy and 960,000 people died as a result of the disease [1]. CT scan is a prominent technique for surgical planning and to diagnose organs in the abdomen [2]. Therefore, CT scan is also regularly used to diagnose liver cancer. Liver segmentation is a critical step in computer-aided therapeutic interpolation using CT images, such as volume estimation, radiation, and surgery of liver transplantation. Manual allocation of each slice is a quiet regular clinical rehearsal for the liver description. As a result, manual segmentation is inefficient, independent, and time-consuming. In this manner, creating a fully automated system capable of diagnosing, monitoring, and expediting therapeutic planning is critical. Numerous techniques for segmenting the liver in CT scans have been described, and a summary of these techniques is provided in [3]. Generally, these methods are classified into three classes, which are interactive [4], semiautomatic [5–7], and automatic [8,9]. Interactive and semiautomatic methods depend on a little or a massive user interaction while automatic methods do not depend on any type of user interaction. Semiautomatic approaches have a potential to reduce the effectiveness of a physician.
To date, there are two types of automatic methods frequently used for liver segmentation. Some approaches based on learning and some approaches are based on antilearning including graph cuts [6,10,11], level set [12,13], and region growing [14]. For some challenging datasets, Wang et al. [15] presented a method that defined a shape-prior-level set method to describe the borders of the liver have a level of changeability that is comparable to those which manually achieved. The popular graph cut approaches for the segmentation of the liver are the extension of standard graph cut method proposed by Boyokov et al. [16,17]. Linguraru et al. [18] presented a method for liver segmentation where the generic-affine-invariant shape parametrization method is integrated with the geodesic-active-contour method and segmented a tumor with the graph cut method. The deformable graph cut method is proposed by Li et al. [19] to detect the surface of the liver using the graph cut with a shape constraint. Rusku et al. [14] proposed a neighborhood-connected region growing method that integrated the local neighborhood of the voxels to segment the 3D liver from CT images. In liver segmentation, level set methods are very successful because it can capture complex shapes and also control shape regularity. Using a sparse representation for local and global information of an image, a level set method is proposed by Sheikhli et al. [20]. Atlas-based methods and active shape models (ASMs) are the traditional learning-based methods. Basically, ASM methods first build a shape of the liver and then tie it to the target image. To challenge the task of liver segmentation, Heimann et al. [21] suggested a technique which is the mixture of SSM and a deformable mesh. A probabilistic active contour method is proposed by Wimmer et al. [22], which combined the shape, boundary and region information into an individual level set equation. A multitiered statistical shape model for liver segmentation is proposed by Erdt et al. [23]. For the precise segmentation of hepatic veins, liver, portal veins, and its tumors simultaneously, Wang et al. [24] presented a method using the sparse-shape-composition (SSC) to build a dynamic contour preceding for the liver. The abovementioned active shape model (ASM) performs well on the liver segmentation but it needs a larger dataset and it is also very complicated to build these time-consuming methods. Then registration algorithms are also needed for the correspondence between the liver atlases and their targets. So, these methods are not independent, they need prior knowledge to tackle the segmentation problems. The registration process is also time-consuming and very complicated in terms of liver abdominal CT images. In atlas-based methods, label fusion and selection of atlases are not easy tasks; therefore, these methods have a very limited clinical usage. When the above machine learning methods in the literature are used on challenging cases that have some limitations, especially in liver segmentation which have an irregular shape, the same intensity level in nearby organs (spleen, heart, and right kidney) and tumor in the liver make it more challenging. Figure 1 shows some of challenges in the CT abdominal images having liver and surrounding organs.
Figure 1.

Sketch of challenges in automatic liver segmentation. The shape of the liver changes in different cases. Nearby organs have a similar intensity level.
Recently, deep learning approaches have earned considerable interest for their ability to learn a hierarchy of features from high to low [36–39]. A review of different deep learning method for COVID-19 is presented [40–42]. These are very powerful methods as compared to machine learning [43], which can acquire more classified features for image segmentation [44] and classification tasks [45–49]. Qadri et al. [25,26], employing a deep belief network, suggested an approach for segmenting the spine from CT images. Ahmad et al. [27] suggested a system for segmenting the liver using stacked autoencoders that learned unsupervised features. A work based on stacked sparse autoencoder (SSAE) for pedestrian gender recognition achieved outstanding results [28]. Hirra et al. [29] used a deep belief network for breast cancer classification using histopathological images. The skin is segmented using a stacked autoencoder to capture the high-level features [50]. CNN is a form of traditional deep learning technique, which may detect nonlinear mappings between its inputs and outputs [51,52] and visual object recognition [53]. Zhang et al. proposed a two-dimensional convolutional neural network for multimodal infant brain image segmentation [31]. In [54], using a combination of deep learning and graph cut techniques, created a fully automatic system for liver segmentation. It starts with a liver segmentation simultaneously using a probability map learned from a 3D CNN. The graph cut method is used to refine the initial segmentation, and this technique is completely automated and capable of simultaneously learning low-to-high-level features. While [32] presented a deep learning technique called a 3D CNN for automated liver segmentation, which trained and got the subject-specific probability map of the liver that acts as a shape prior and gives the initial surface of the liver. This model is built on the preceding segmentation using local and global information. The global information is used in the healthy part of the liver to learn the area appearance and intensity dissemination, and the local nonparametric information is used to get the abnormal liver information. To resolve the problem of liver segmentation [33], a new 3D convolutional neural network is presented. The suggested model 3D-DSN is better than the simple CNN model in terms of discrimination capability, optimization, efficiency, and effectiveness. 3D-DSN model is a fully convolutional model that is effective in learning. The deep supervision approach to the hidden layer in this work has been used which improve the convergence rate of optimization and increase the precision. To refine the segmentation in postprocessing, a conditional random field is used. Liver and its tumor segmentation are carried out using Octave CNN, which got better results [35,55]. Deep learning methods are being used to detect the tumor from CT images and get reasonable results [56,57]. FCNN was trained on a 3D CT volume of the liver [34], and then a cascade fully convolutional neural network (CFCN) was applied to CT data slices to extract the liver and its lesion. Afterward, 3D-CRF is applied for postprocessing to enhance the segmentation results. Larger parameters, sufficient memory requirements, and extra computational powers are needed in fully 3D convolutional neural network models [58]. To avoid the fully three-dimensional CNN network, two-dimensional patches are utilized and processed these patches for segmentation. The major reason behind it is to discourage the fully 3D CNN because this method is very much expensive in terms of parametric complexity which takes longer time to train the network. So, over segmentation problem might occur due to the biases in rare classes [59]. Table 1 summarizes the recently published deep learning methods. There are some recent deep learning methods which have been used for the detection of different diseases that performed well [60–67].
Table 1.
The recently published deep learning methods to perform different classification and segmentation tasks.
| S. no. | Author | Technique name | Purpose | Date published |
|---|---|---|---|---|
| 1 | Qadri et al. [25] | Deep belief network with two hidden layers | Deep belief network is used for the spine segmentation from CT images. | 2018 |
| 2 | Qadri et al. [26] | Pa-DBN with two hidden layers | Employing a deep belief network, suggested an approach for segmenting the spine from CT images. | 2019 |
| 3 | Ahmad et al. [27] | DSAE with two hidden layers | Segmenting the liver from CT images using stacked autoencoders that learned unsupervised features. | 2017 |
| 4 | Raza et al. [28] | Stacked sparse autoencoder (SSAE) | HOG-based features generated and then sparse stacked autoencoder is used for pedestrian gender recognition. | 2018 |
| 5 | Hirra et al. [29] | Pa-DBN-BC | Deep belief network for breast cancer classification using histopathological images. | 2021 |
| 6 | Lei et al. [30] | Stacked autoencoder | The skin is segmented using a stacked autoencoder to capture the high-level features. | 2016 |
| 7 | Zhang et al. [31] | CNN with three convolutional and a fully connected layer is used; after the convolutional layer, local response normalization layer is used | Propose a two-dimensional convolutional neural network for multimodal infant brain image segmentation. | 2015 |
| 8 | Hu et al. [32] | 3D CNN | Presented a deep learning technique called a 3D CNN for automated liver segmentation which trained and got the subject-specific probability map of the liver that acts as a shape prior and gives the initial surface of the liver. | 2016 |
| 9 | Dou et al. [33] | 3D-DSN | The suggested model 3D-DSN is better than the simple CNN model in terms of discrimination capability, optimization, efficiency, and effectiveness. | 2016 |
| 10 | Christ. et al. [34] | FCNN | FCNN is trained on a 3D CT volume of the liver and then a cascade fully convolutional neural network (CFCN) was applied to CT data slices to extract the liver and its lesion. Afterward, 3D-CRF is applied for postprocessing to enhance the segmentation results. | 2017 |
| 11 | Ahmad, et al. [35] | DBN-DNN | Liver segmentation is performed from CT images using deep belief network. This method gives good accuracy and DSC. | 2019 |
We extracted the 2D patches from our datasets and processed them with a 2D convolutional neural network to avoid the complexity of fully 3D CNN. Preprocessing is performed to get the enhanced CT image. The input of 2D patches is given to CNN then acquire the probability map:
We developed a lightweight convolutional neural network to separate the liver from CT scan images.
The proposed CNN model consists of three convolutional and two fully connected layers hence efficiently segmenting the liver.
The weights are initialized using a random Gaussian distribution, which does a distance preserving of the information. We added the LRN layer after each convolutional layer to enhance the data adaptability.
Three benchmark datasets SLiver'07, 3Dircadb01, and LiTS17 are used to analyze the model performance which improves the stability of our model.
The whole article is organized as follows: Sections 2–4 are the proposed method, experimental setup, and results and discussion, respectively. The conclusion is given in Section 5.
2. Proposed Method
This section provides details of the method we followed in this study. The whole pipeline of our system is illustrated in Figure 2.
Figure 2.

Flow diagram of our methodology.
2.1. Data Preparation
A simple preprocessing is applied in the proposed work on each axial CT slice. We enhanced the contrast and normalized the dataset using zero mean and unit variance to obtain the data range between 0 and 1. On both the training and testing datasets, we applied similar steps. For the training dataset, we applied the additional step to augment the input data. We crop the train data where liver slices are not affected and rotate the images on 90, 180, and 270°. The detail of data augmentation is given in the literature [68,69]. We sliced each axial 2D image into 32 × 32 patches for both training and testing data and picked 1.3 million patches from the background and foreground randomly for training data. The image patches of two classes are equally taken with a ratio of 1 : 1 for stable training and validation.
2.2. Convolutional Neural Network
We briefly discuss the convolutional neural network (CNN) in this section. More detail on CNN can be found in [51,70]. The CNN is a discrepancy of a multilayer perceptron. The core part of CNN is based on convolutional and subsampling layers. For the hierarchy of feature learning in the convolutional neural network, numerous convolutional layers are heaped together. Every convolutional layer is capable of obtaining the feature map from the layer before it. These layers are connected with some filters. We indicate that Con(n − 1) and Con(n) are the input and output of the n convolutional layers, respectively, and Coni(n) denotes ith feature map of the nth layer in a particular CNN. The following equation can be used to calculate the output of the nth layer.
| (1) |
Here, Wi,j(n) is the kernel relating of ith and jth output map and ∗ is the denotation of the convolution and bjn denotes the bias of the jth output record of the nth layer in a particular convolutional layer. FW, b(.) denotes a nonlinear activation function. For nonlinear activation, the function has many options, which are rectified linear unit (ReLU), sigmoid, and hyperbolic tangent. A subsampling layer after the convolutional layer is used to reduce the computational complexity. So, in this manner, we adapt the max-pooling layer which is commonly used. Fully connected layers are often used after the convolutional layers and at the end, a softmax classifier is used to classify the results. In binary classification problems, the regression model is being used to normalize the results of kernel convolution. Convolutional layers are sharing the weights which is the key advantage of CNN. Each pixel in the layer uses the same filter. The main advantage is to improve the performance as well as reduce the memory size.
If the training set is made up of m labeled samples,
{(x1, y1), (x2, y2),…, (xm, ym) }, yi can be 0 or 1, where I = (1, 2, 3,…, m).
θ indicates all the important parameters including bias, kernel, and softmax parameters of the CNN. Concerning θ, we reduce the following cost function for logistic regression.
| (2) |
For regularization of classification, weight decay can be used to castigate the larger values of the softmax layer. Gradient-based optimization can be utilized to reduce the cost function [52] and backpropagation is used to calculate the partial derivatives [51].
2.3. Proposed Ga-CNN Architecture
For any low-dimensional data, random estimation is a general sampling approach [71–73]; a deep neural network with random weights is a general system that can separate the data. The angles among its points, it is observed that there are larger angles among different classes [74–77]. The intraclass angles are more important than interclass ones in a deep network.
The recent study shows that the vital property of Gaussian distribution is to preserve the ratio among the angles in the data during the DNN training. Therefore, it does not give a priority in any direction but it treats the same to each direction; the above discussion is leading to the behavior of distances and angles throughout the DNN net. In common, there are small angles between the points in the same class but have a larger distance in different classes. Random Gaussian weights would be a good choice for the DNN network parameters if it holds all the points. We adopted the random Gaussian weights for the initialization of the training network, where the hypothetical details and proofs are given in the literature [78]. To address the problem of liver segmentation in CT images, a patch-based methodology is being developed. The first 2D image is separated into M number of patches. The input image of axial slice with size of 512 × 512 is divided into 1 × 32 × 32 patch size, where 1 is the channel of the image patch. Figure 3 shows the foreground and background patches.
Figure 3.

The foreground and background patches. Each patch size is 32 × 32 pixels, and M is the total number of extracted patches.
The proposed DNN method has the input of 1 × N × N 2D patch of CT scan images. The most important and necessary layer for building the CNN model is a convolutional layer. Numerous convolutional layers are placed on top of one another to create a feature hierarchy. Each layer receives its input from the layer that was previously connected. The convolutional layer executes the stack of input CT image patches and outputs the number of feature maps. Each region of interest corresponds to a topological arrangement inside the responsive map of a specific nonlinear spatial feature abstraction. These learning parameters could very certainly be applied to any spatial neighborhood layout using the sliding window architecture. The convolutional layer activation process is dependent on the single plan of an input CT image. The feature map from previous layers is often included in the information plan for subsequent layers. In the proposed model, we are using three convolutional and two fully connected layers. Convolutional layers are the main components of CNN, reducing the number of dimensions while increasing the depth. The filter size of 7 × 7 in the first, 5 × 5 in the second, and 3 × 3 in the third convolutional layer are used. Figure 4 shows the proposed Ga-CNN network architecture. The other information regarding each layer in our model is given in Table 2.
Figure 4.

Proposed Ga-CNN architecture.
Table 2.
The CNN model's detailed architecture is employed in this work.
| Type of layer | Kernel | Stride | Padding | Output | Depth | Trainable parameters |
|---|---|---|---|---|---|---|
| Input | — | — | — | 32 × 32 | 1 | |
| Convolution | 7 | 2 | 0 | 13 × 13 | 32 | 1600 |
| Max-Pooling | 3 | 2 | 0 | 13 × 13 | 32 | 0 |
| Convolution | 5 | 1 | 2 | 13 × 13 | 192 | 153792 |
| Max-Pooling | 3 | 2 | 0 | 6 × 6 | 192 | 0 |
| Convolution | 3 | 1 | 1 | 6 × 6 | 256 | 442624 |
| Max-Pooling | 2 | 2 | 0 | 3 × 3 | 256 | 0 |
| Fully connected | 1 | — | — | 1 × 1 | 4096 | 9,441,280 |
| Fully connected | 1 | — | — | 1 × 1 | 2 | 8,194 |
| Softmax | — | — | — | 1 × 1 | 2 | |
| Total | 10,047,490 |
Each convolutional layer is tailed by a ReLU, LRN, and max-pooling layer. Our network contains three max-pooling layers, where 3 × 3 sub-window size is being used with stride of 2 in the first and second max-pooling layers and a sub-window size of 2 × 2 with a stride of 2 is used in every path of third max-pooling layer [79]. ReLU is a piecewise linear transformation max (x, 0) that improves convergence by selecting important invariant features that enhance the execution's generality [51]. ReLU has a required property that does not need any input normalization to stop them from soaking. If a positive input is produced by a training example that is given to ReLU, that neuron will be learned. Therefore, LRN still adds the generalization. The following formula is used to calculate the LRN.
| (3) |
where bx,yi is the response normalization activity, where the sum runs across N neighboring kernel maps and the layer comprises N number of kernels at the same spatial position. We used the same LRN hyperparameters as in [51]. The convolutional layer produces the output of the feature mapping procedure as an input from the preceding convolutional layer. From the perspective of the neural network, feature maps are associated with the hidden layers of neurons. Each coordinate subset of features is associated with a single neuron, and the size of the receptive field is proportional to the kernel size. Finally, at the network's end, two fully connected layers are added. The fully connected layer executes the scalar product, where the input is a vector and it returns a scalar. The first fully connected layer with a neuron size of 4096 and the dropout factor of 0.3 is used. Dropout is employed during the training phase to keep the network from overfitting. The second fully connected layer which is final layer of our system, and then softmax classifier is used to perform classification for our model (Background and Liver). We reconstructed the image on the bases of classification results to perform a segmentation. Meanwhile, after segmenting each 2D liver image, we used a 5 × 5 median filter to smother the resultant liver. Total trainable parameters of our model (Ga-CNN) are about 10 million; if we compare these parameters with Alexnet, which has 62 million, and VGG16, they are about 138 million. In terms of complexity, our model is less complex and stable. The information of trainable parameters is given in Table 2.
3. Experimental Setup
In total, 60 contrast-enhanced CT images with ground truths were used to train the network. We used three challenging datasets in our work, which are the SLiver'07 training set, 3Dircadb01, and LiTS17 training dataset. We have selected 10 CT images from the SLiver'07 training set, 10 CT images from 3Dircadb01, and 40 CT images from the LiTS dataset for training and validation. All three benchmark datasets are freely available online. In total, 30 contrast-enhanced CT images were selected for testing purposes, where the axial slice resolution was 512 × 512 pixels. We randomly selected 10 CT images from SLiver'07, 10 CT images from 3Dircadb01, and 10 CT images from the LiTS17 dataset. For experiments, we adopted the MATLAB 2021a for implementation of our model. We selected the parameters which are Stochastic Gradient Descent with Momentum (SGDM) and initial learning rate of 0.01 for this experiment. A computer with windows operating system, Intel Core i-7, 8565U processor, 32 GB of RAM with 2 GB GPU (NVIDEA GeForce 250) was used.
In our method, we initialized the weights manually from a Gaussian distribution with a standard deviation of 0.0001. We applied Gaussian weights to CNN network which perform steady inserting of the input data that draw a linking among the features produced by our model [78]. We trained our model with 70 epochs where we set the initial learning rate of 0.01. After 20 epochs, the learning rate dropped by a factor of 0.1. Momentum was constant in the whole experiment, which was 0.9. Weight decay and mini-batch size were 0.0001 and 64, respectively. Stochastic gradient descent with momentum (SGDM) is used as an optimization algorithm. The SGD algorithm has trouble, where the surface curve is much tighter in one dimension than another [80]. Momentum is the method that can help the SGD to accelerate in a relevant direction [81]. The training hyperparameters are given in Tables 3 and 4. The proposed model performance is analyzed by key modules that affect and extract the patches. Figure 5 shows the visualization of three convolutional layers in our training model.
Table 3.
Hyperparameters of Ga-CNN.
| Parameter | Value |
|---|---|
| Momentum | 0.9 |
| L2 Regularization | 0.0001 |
| Max epochs | 70 |
| Mini-batch size | 64 |
Table 4.
Learning rate schedule.
| Learning rate | Value |
|---|---|
| Initial | 0.01 |
| Schedule | Piecewise |
| Drop factor | 0.1 |
| Drop period | 20 |
Figure 5.

Visualization results of filters of each convolutional layer. The 1st convolutional layer with 32 filters, the 2nd convolutional layer with 192 filters, and the 3rd convolutional layer with 256 filters.
After the training, the trained CNN algorithm is learned iteratively from the initial liver probability map. One of the test volumes in Figure 6 shows the iterative results of the liver probability map. In the 8th iteration, the liver and spleen had the same intensity and texture level but in the 15th iteration, the liver becomes brighter which means liver pixels are stronger and easily differentiable. In a 25th iteration, the spleen becomes smaller as in previous iterations. In the 35th iteration, the spleen disappeared and the liver becomes brighter. In the 60th to 70th iterations, the validation results converge and the shape of the liver becomes accurate.
Figure 6.

Iterative results of liver probability map produced by Ga-CNN model on CT scan image of SLiver'07 dataset. From top left to bottom right, 8th, 15th, 25th, 55th, 60th, and 70th iterative liver probability maps are shown. The brighter region shows the more probability map of the liver.
4. Results and Discussion
In this section, we provide the detailed segmentation results based on three benchmark datasets which are SLiver'07, 3Dircadb01, and LiTS17. The input image is converted into multiple 2D patches and given as an input to Ga-CNN model for detection and then segmentation is performed on the input image.
All the three benchmark datasets (having normal and abnormal CT scan images) are evaluated on eight performance metrics such as SE, SP, ACC, Precision, FPR, FNR, JSI, and DSC. In equations 4 to 11, true positive (TP) denotes all the pixels related to the liver, true negative (TN) shows all the pixels related to the background. False negative (FN) denotes the liver pixels that do not classify accurately, and false positive (FP) denotes the pixels which belong to the background but are not identified as a background. The dice similarity coefficient (DSC) is being used for the calculation of automatic and manual segmentation, and JSI is used to measure the similarity between the two images.
| (4) |
| (5) |
| (6) |
| (7) |
| (8) |
| (9) |
| (10) |
| (11) |
The analysis of key components is analyzed which affect and extract patches on the proposed model. In this section, previous deep learning and machine learning methods are being compared. Finally, the outcomes of three benchmark datasets are conveyed by experimentation.
We presented the results of 10 CT images that were randomly selected out of 20 CT images for testing purpose in Sliver'07 dataset. The overall DSC of 95.0%, Jaccard similarity index of 90.47%, the accuracy of the proposed model of 95.1%, precision of 97.2%, sensitivity of 95%, specificity of 95.2%, FNR of 0.048, and FPR of 0.05 are noted on the SLiver'07 dataset. We observed from the above results that our method performed well in classification as well as in segmentation. Figure 7 depicts the segmentation results on 2D axial slices of the SLiver'07 dataset on the proposed model. The axial slices are randomly chosen from different CT images of the SLiver'07 dataset. Segmentation results show that the proposed model efficiently recognized the boundaries of the liver, where (a) is the original CT image, (b) is the original label, (c) is the segmentation results of Ga-CNN model, and (d) is the overlapping results of the original label (green) and segmented results of Ga-CNN (red).
Figure 7.

The results of segmentation of SLiver'07 dataset.
The qualitative results of 3Dircadb01 datasets are presented in Table 6. 3Dircadb01 dataset composed of 20 CT images; we presented the mean results of 10 CT images, which were selected from 3Dircadb01 dataset randomly for testing. The DSC of 92.9%, JSI of 85.1%, the accuracy of the proposed model of 93.1%, precision of 95.5%, sensitivity of 93.0%, specificity of 93.2%, FNR of 0.08, and FPR of 0.07 are noted on the 3Dircadb01 dataset. Figure 8 shows the segmentation results on 2D axial slices of the former dataset on the proposed model. The axial slices are randomly chosen from different CT images of the 3Dircadb01 dataset, where (a) is the original CT image, (b) is the original label, (c) is the segmentation results of Ga-CNN model, and (d) is the overlapping results of original label (green) and segmented results of Ga-CNN (red).
Table 6.
The segmentation results of the 3Dircadb01 dataset using the proposed model.
| Dataset | DSC% | JSI% | ACC% | Precision% | SE% | SP% | FNR | FPR |
|---|---|---|---|---|---|---|---|---|
| 3Dircadb01 | 92.9 | 86.74 | 93.1 | 95.5 | 93.0 | 93.2 | 0.08 | 0.07 |
Figure 8.

The results of segmentation of 3Dircadb01 dataset.
We presented the mean results of 10 CT images which were selected from LiTS dataset randomly in Table 7. The DSC of 97.31%, JSI of 94.76%, the accuracy of the proposed model of 97.25%, precision of 97.06%, sensitivity of 97.56%, specificity of 96.93%, FNR of 0.02, and FPR of 0.02 are noted on LiTS dataset. The axial slices are randomly chosen from different CT images of LiTS dataset. Figure 9 shows the results of LiTS17 dataset, where (a) is the original CT image, (b) is the original label, (c) is the segmentation results of Ga-CNN model, and (d) is the overlapping results of original label (green) and segmented results of Ga-CNN (red).
Table 7.
The segmentation results of LiTS17 training dataset using the proposed model.
| Dataset | DSC% | JSI% | ACC% | Precision% | SE% | SP% | FNR | FPR |
|---|---|---|---|---|---|---|---|---|
| LiTS17 | 97.31 | 94.76 | 97.25 | 97.06 | 97.56 | 96.93 | 0.02 | 0.03 |
Figure 9.

The results of segmentation of LiTS dataset.
The intermediate results of all the three datasets are given in Table 8. The mean DSC of 95.07%, JSI of 90.65%, ACC of 95.15%, precision of 96.58%, SE of 95.18%, SP of 95.11%, FNR of 0.05, and FPR of 0.05 are noted. The overall performance of the proposed model is better and can be comparable to another state-of-the-art methods in the literature. Table 9 depicts the comparison of the proposed method results with other recently published well-known liver segmentation methods.
Table 8.
The intermediate results of the proposed model on three benchmark datasets.
| Dataset | DSC% | JSI% | ACC% | Precision% | SE% | SP% | FNR | FPR |
|---|---|---|---|---|---|---|---|---|
| SLiver'07 | 95.00 | 90.47 | 95.10 | 97.20 | 95.00 | 95.20 | 0.048 | 0.05 |
| 3Dircadb01 | 92.9 | 86.74 | 93.10 | 95.50 | 93.00 | 93.20 | 0.08 | 0.07 |
| LiTS17 | 97.31 | 94.76 | 97.25 | 97.06 | 97.56 | 96.93 | 0.02 | 0.03 |
| Mean | 95.07 | 90.65 | 95.15 | 96.58 | 95.18 | 95.11 | 0.05 | 0.05 |
Table 9.
The comparison with other methods.
The comparison with other recently published liver segmentation methods is given in this section. In [82], liver segmentation is performed on CT images using two challenging datasets which are 3Dircadb01 and SLiver'07. DSC of SLiver'07 and 3Dircadb01 datasets are 94.03% and 91.19%, respectively. This method is somewhat complex and needs to remove the rib from the CT image and then apply the random walker algorithm to extract the liver but the proposed method do not need to remove any other organ first. Ga-CNN detects the liver boundaries very efficiently without any other postprocessing support. If we compare our method with that in [35], the DSC of our method is 95.07% which is better. We conducted the experiments on three benchmark datasets with 30 contrast-enhanced CT images, that can also validate our algorithm. Another recently published deep learning method using a stacked autoencoder (DSAE) [27] got 90.1% DSC that is less than the proposed method. The training and validation of Deep stacked auto encoder (DSAE) are performed on 2 publicly available datasets (3Dircadb and Sliver'07); the limitations of this method against our methods are obvious that DSAE cannot detect the abnormal liver in the image very well but our method can efficiently segment the abnormal liver very efficiently. In [46], the DSC of liver segmentation was noted 92% using LiTS17 dataset that is less than the DSC in our proposed work. In [35], the two datasets, SLiver'07 and 3Dircadb01, that were used to segment the liver show the DSC of 94.80% and 91.83%, respectively, which are obviously much fewer than Ga-CNN method. Proposed algorithm has a capability to find the edges and curves very accurately, and the computational time for each slice is only 6 seconds, which shows that Ga-CNN is very lightweight. The application of our method is best for the donors of the liver, because the normal liver can be efficiently segmented with correct boundaries. From the segmentation results shown in Figures 7–9, it is clearly observed that the proposed method correctly segments the areas of the liver from the CT images, where the heart, stomach and spleen also having the same intensity level. Normally, the spleen and stomach are mis-segmented by many algorithms because of their same intensity nature, but our DNN correctly identifies the liver pixels and discriminates the other organs due to the efficient feature learning. Liver shape variability is also a big problem in CT images but our algorithm can solve it effectively.
In the above discussion, we compared our method with deep learning and some other liver segmentation methods. We observed the above given experimental results that the proposed technique is better as compared to the other recent methods in the literature. This proves the applicability of the proposed method. The validation results of three well-known benchmark datasets also make our algorithm very proficient. Overall, we found in our research that initialization of DNN with random Gaussian weights perform better and have a steady insertion of a data. The most important is to draw the connection among the dimensions that are generated by the DNN and still retain the complication of the data and matric information of the original manifold. The results of the proposed method show that deep neural networks are general classifier for that data which is based on the angles of major alliance among the classes in the data. Our random Gaussian initialization is mainly focused on the previous study that proved that this strategy is also better to train the DNN models with low training samples. For liver segmentation, this also helps us improve the results of the abnormal liver, where the angles and shapes of the liver could be detected in a better way. The 3Dircadb01 and LiTS are both complex datasets, having different types of tumors in it. But the mean DSC and precision of these two datasets prove that our proposed model performed well and can be approachable against other well-known methods.
5. Conclusion and Future Work
This study proved the development of an accurate and robust technique of a liver segmentation. The developed algorithm accomplished the segmentation with a very light convolutional neural network (Ga-CNN) with randomly initialized Gaussian weights, which performed a steady insertion of the data that draw the linking between the magnitudes of the features produced by our network that still preserve the metric evidence of the complexity of the data and the original manifold. Adding ReLU activation function after each convolutional layer makes the learning very fast and found early convergence. Also, we introduced local response normalization (LRN) after ReLU which improved the precision and reduced the false positive rate. The validation of three benchmark datasets including SLiver'07, 3Dircadb01, and LiTS17 prove that the proposed model performed better on CT images to segment the liver. The recorded mean dice, Jaccard index, accuracy, and precision were 95.07%, 90.65%, 95.15%, and 96.58%, respectively, on 30 CT test images.
We tested our model on three datasets, which gives better accuracy and DSC. On the bases of these results, we are sure that our lightweight CNN could give better results for the segmentation of liver with other datasets. In future, we will add more dataset for more accuracy. Moreover, our method could be applied on other organs for segmentation, like the kidney, lungs, and heart. Our lightweight CNN model is limited for the segmentation of a liver. On complex datasets like 3Dircadb01, the accuracy of Ga-CNN is less than other used datasets. For this purpose, we will improve our preprocessing techniques to cope this situation. We will also focus on liver tumor segmentation using Ga-CNN with more improvements.
Table 5.
The segmentation results of SLiver'07 training dataset using the proposed model.
| Dataset | DSC% | JSI% | ACC% | Precision% | SE% | SP% | FNR | FPR |
|---|---|---|---|---|---|---|---|---|
| SLiver'07 | 95.0 | 90.47 | 95.1 | 97.2 | 95 | 95.2 | 0.048 | 0.05 |
Acknowledgments
This research was supported by the Researchers Supporting Project (RSP2022R476), King Saud University, Riyadh, Saudi Arabia.
Contributor Information
Mubashir Ahmad, Email: mubashir_bit@yahoo.com.
Syed Furqan Qadri, Email: furqangillani79@gmail.com.
Hayat Mansoor Alaghbari, Email: hayat.alaghbari@taiz.edu.ye.
Data Availability
The data supporting the findings of this study are included within the article.
Conflicts of Interest
The authors declare that there are no conflicts of interest.
Authors' Contributions
Mubashir Ahmad and Syed Furqan Qadri contributed equally as first authors. All the authors contributed to acquisition, analysis, or of data.
References
- 1.Jemal A., Bray F., Center M. M., Ferlay J., Ward E., Forman D. Global cancer statistics. CA: A Cancer Journal for Clinicians . 2011;61(2):69–90. doi: 10.3322/caac.20107. [DOI] [PubMed] [Google Scholar]
- 2.Selvalakshmi V., Devi S. N. Segmentation and 3D visualization of liver, lesions and major blood vessels in abdomen CTA images. Biomedical Research . 2017;28(16):7206–7212. [Google Scholar]
- 3.Heimann T., van Ginneken B., Styner M. A. Comparison and evaluation of methods for liver segmentation from CT datasets. IEEE Transactions on Medical Imaging . 2009;28(8):1251–1265. doi: 10.1109/tmi.2009.2013851. [DOI] [PubMed] [Google Scholar]
- 4.Beichel R., Bornik A., Bauer C., Sorantin E. Liver segmentation in contrast enhanced CT data using graph cuts and interactive 3D segmentation refinement methods. Medical Physics . 2012;39(3):1361–1373. doi: 10.1118/1.3682171. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Freiman M., Eliassaf O., Taieb Y., Joskowicz L., Azraq Y., Sosna J. An iterative Bayesian approach for nearly automatic liver segmentation: algorithm and validation. International Journal of Computer Assisted Radiology and Surgery . 2008;3(5):439–446. doi: 10.1007/s11548-008-0254-1. [DOI] [Google Scholar]
- 6.Peng J., Dong F., Chen Y., Kong D. A region-appearance-based adaptive variational model for 3D liver segmentation. Medical Physics . 2014;41(4) doi: 10.1118/1.4866837.043502 [DOI] [PubMed] [Google Scholar]
- 7.Peng J., Wang Y., Kong D. Liver segmentation with constrained convex variational model. Pattern Recognition Letters . 2014;43:81–88. doi: 10.1016/j.patrec.2013.07.010. [DOI] [Google Scholar]
- 8.Massoptier L., Casciaro S. Fully automatic liver segmentation through graph-cut technique. Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2007 (EMBS 2007); August 2007; Lyon, France. pp. 5243–5246. [DOI] [PubMed] [Google Scholar]
- 9.Tomoshige S., Oost E., Shimizu A., Watanabe H., Nawano S. A conditional statistical shape model with integrated error estimation of the conditions; application to liver segmentation in non-contrast CT images. Medical Image Analysis . 2014;18(1):130–143. doi: 10.1016/j.media.2013.10.003. [DOI] [PubMed] [Google Scholar]
- 10.Afifi A., Nakaguchi T. Liver segmentation approach using graph cuts and iteratively estimated shape and intensity constrains. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; October 2012; Nice, France. pp. 395–403. [DOI] [PubMed] [Google Scholar]
- 11.Chen X., Bagci U. 3D automatic anatomy segmentation based on iterative graph-cut-ASM. Medical Physics . 2011;38(8):4610–4622. doi: 10.1118/1.3602070. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Lee J., Kim N., Lee H., et al. Efficient liver segmentation using a level-set method with optimal detection of the initial liver boundary from level-set speed images. Computer Methods and Programs in Biomedicine . 2007;88(1):26–38. doi: 10.1016/j.cmpb.2007.07.005. [DOI] [PubMed] [Google Scholar]
- 13.Pan S., Dawant B. M. Automatic 3D segmentation of the liver from abdominal CT images: a level-set approach. Proceedings of the Medical Imaging 2001: Image Processing; July 2001; San Diego, CA, USA. pp. 128–139. [Google Scholar]
- 14.Ruskó L., Bekes G., Fidrich M. Automatic segmentation of the liver from multi- and single-phase contrast-enhanced CT images. Medical Image Analysis . 2009;13(6):871–882. doi: 10.1016/j.media.2009.07.009. [DOI] [PubMed] [Google Scholar]
- 15.Wang J., Cheng Y., Guo C., Wang Y., Tamura S. Shape-intensity prior level set combining probabilistic atlas and probability map constrains for automatic liver segmentation from abdominal CT images. International journal of computer assisted radiology and surgery . 2016;11(5):817–826. doi: 10.1007/s11548-015-1332-9. [DOI] [PubMed] [Google Scholar]
- 16.Boykov Y., Funka-Lea G. Graph cuts and efficient N-D image segmentation. International Journal of Computer Vision . 2006;70(2):109–131. doi: 10.1007/s11263-006-7934-5. [DOI] [Google Scholar]
- 17.Boykov Y. Y., Jolly M.-P. Interactive graph cuts for optimal boundary & region segmentation of objects in ND im ages. Proceedings of the Eighth IEEE International Conference on Computer Vision. ICCV 2001; July 2001; Vancouver, BC, Canada. pp. 105–112. [Google Scholar]
- 18.Linguraru M. G., Richbourg W. J., Watt J. M., Pamulapati V., Summers R. M. Liver and tumor segmentation and analysis from CT of diseased patients via a generic affine invariant shape parameterization and graph cuts. Proceedings of the International MICCAI Workshop on Computational and Clinical Challenges in Abdominal Imaging; September 2011; Bethesda, MD, USA. pp. 198–206. [Google Scholar]
- 19.Li G., Chen X., Shi F., Zhu W., Tian J., Xiang D. Automatic liver segmentation based on shape constraints and deformable graph cut in CT images. IEEE Transactions on Image Processing . 2015;24(12):5315–5329. doi: 10.1109/tip.2015.2481326. [DOI] [PubMed] [Google Scholar]
- 20.Al-Shaikhli S. D. S., Yang M. Y., Rosenhahn B. Automatic 3D liver segmentation using sparse representation of global and local image information via level set formulation. https://arxiv.org/abs/1508.01521, 2015.
- 21.Gauriau R., Cuingnet R., Prevost R., et al. A generic, robust and fully-automatic workflow for 3D CT liver segmentation. Proceedings of the International MICCAI Workshop on Computational and Clinical Challenges in Abdominal Imaging; September 2013; Nagoya, Japan. pp. 241–250. [Google Scholar]
- 22.Wimmer A., Soza G., Hornegger J. A generic probabilistic active shape model for organ segmentation. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; September 2009; London, UK. pp. 26–33. [DOI] [PubMed] [Google Scholar]
- 23.Qadri S. F., Ahmad M., Ai D., Yang J., Wang Y. Deep belief network based vertebra segmentation for CT images. Proceedings of the Chinese Conference on Image and Graphics Technologies; August 2018; Beijing, China. pp. 536–545. [Google Scholar]
- 24.Wang G., Zhang S., Li F., Gu L. A new segmentation framework based on sparse shape composition in liver surgery planning system. Medical Physics . 2013;40(5) doi: 10.1118/1.4802215.051913 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Qadri S. F., Ahmad M., Ai D., Yang J., Wang Y. Deep belief network based vertebra segmentation for CT images. Proceedings of the Chinese Conference on Image and Graphics Technologies; August 2018; pp. 536–545. [Google Scholar]
- 26.Furqan Qadri S., Ai D., Hu G., et al. Automatic deep feature learning via patch-based deep belief network for vertebrae segmentation in CT images. Applied Sciences . 2019;9(1):p. 69. [Google Scholar]
- 27.Ahmad M., Yang J., Ai D., Qadri S. F., Wang Y. Deep-stacked auto encoder for liver segmentation. Proceedings of the Chinese Conference on Image and Graphics Technologies; November 2017; Beijing, China. pp. 243–251. [Google Scholar]
- 28.Raza M., Sharif M., Yasmin M., Khan M. A., Saba T., Fernandes S. L. Appearance based pedestrians’ gender recognition by employing stacked auto encoders in deep learning. Future Generation Computer Systems . 2018;88:28–39. doi: 10.1016/j.future.2018.05.002. [DOI] [Google Scholar]
- 29.Hirra I., Ahmad M., Hussain A., et al. Breast cancer classification from histopathological images using patch-based deep learning modeling. IEEE Access . 2021;9:24273–24287. doi: 10.1109/access.2021.3056516. [DOI] [Google Scholar]
- 30.Lei Y., Yuan W., Wang H., Wenhu Y., Bo W. A skin segmentation algorithm based on stacked autoencoders. IEEE Transactions on Multimedia . 2016;19(4):740–749. [Google Scholar]
- 31.Zhang W., Li R., Deng H., et al. Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage . 2015;108:214–224. doi: 10.1016/j.neuroimage.2014.12.061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Hu P., Wu F., Peng J., Liang P., Kong D. Automatic 3D liver segmentation based on deep learning and globally optimized surface evolution. Physics in Medicine and Biology . 2016;61(24):8676–8698. doi: 10.1088/1361-6560/61/24/8676. [DOI] [PubMed] [Google Scholar]
- 33.Dou Q., Chen H., Jin Y., Yu L., Qin J., Heng P.-A. 3D deeply supervised network for automatic liver segmentation from CT volumes. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; July 2016; Athens, Greece. pp. 149–157. [Google Scholar]
- 34.Christ P. F., Ettinger F., Grun F., et al. Automatic liver and tumor segmentation of ct and MRI volumes using cascaded fully convolutional neural networks. https://arxiv.org/abs/1702.05970, 2017.
- 35.Ahmad M., Ai D., Xie G., et al. Deep belief network modeling for automatic liver segmentation. IEEE Access . 2019;7:20585–20595. doi: 10.1109/access.2019.2896961. [DOI] [Google Scholar]
- 36.Gumaei A., Sammouda R., Al-Rakhami M., AlSalman H., El-Zaart A. Feature selection with ensemble learning for prostate cancer diagnosis from microarray gene expression. Health Informatics Journal . 2021;27(1) doi: 10.1177/1460458221989402.1460458221989402 [DOI] [PubMed] [Google Scholar]
- 37.Alamri A., Gumaei A., Al-Rakhami M., Hassan M. M., Alhussein M., Fortino G. An effective bio-signal-based driver behavior monitoring system using a generalized deep learning approach. IEEE Access . 2020;8:135037–135049. doi: 10.1109/access.2020.3011003. [DOI] [Google Scholar]
- 38.Al-Rakhami M. S., Islam M. M., Islam M. Z., Asraf A., Sodhro A. H., Ding W. Diagnosis of COVID-19 from X-rays using combined CNN-RNN architecture with transfer learning. MedRxiv . 2020:p. 20181339. [Google Scholar]
- 39.Islam M. An efficient human computer interaction through hand gesture using deep convolutional neural network. SN Computer Science . 2020;1(4):1–9. doi: 10.1007/s42979-020-00223-x. [DOI] [Google Scholar]
- 40.Islam M. M., Karray F., Alhajj R., Zeng J. A review on deep learning techniques for the diagnosis of novel coronavirus (COVID-19) IEEE Access . 2021;9:30551–30572. doi: 10.1109/access.2021.3058537. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Rahman M. M., Manik M. M. H., Islam M. M., Mahmud S., Kim J.-H. An automated system to limit COVID-19 using facial mask detection in smart city network. Proceedings of the IEEE International IOT, Electronics and Mechatronics Conference (IEMTRONICS); October 2020; Vancouver, BC, Canada. pp. 1–5. [Google Scholar]
- 42.Rahman M. M., Islam M., Manik M., Hossen M., Al-Rakhami M. S. Machine learning approaches for tackling novel coronavirus (COVID-19) pandemic. SN Computer Science . 2021;2(5):1–10. doi: 10.1007/s42979-021-00774-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Gumaei A., Hassan M. M., Hassan M. R., Alelaiwi A., Fortino G. A hybrid feature extraction method with regularized extreme learning machine for brain tumor classification. IEEE Access . 2019;7:36266–36273. doi: 10.1109/access.2019.2904145. [DOI] [Google Scholar]
- 44.Gomai A., El-Zaart A., Mathkour H. An efficient iris segmentation approach. Proceedings of the International Conference on Graphic and Image Processing (ICGIP 2011); September 2011; Cairo, Egypt. pp. 444–450. [Google Scholar]
- 45.Havaei M., Davy A., Warde-Farley D., et al. Brain tumor segmentation with deep neural networks. Medical Image Analysis . 2017;35:18–31. doi: 10.1016/j.media.2016.05.004. [DOI] [PubMed] [Google Scholar]
- 46.Ma J., Deng Y., Ma Z., Mao K., Chen Y. A liver segmentation method based on the fusion of VNet and WGAN. Computational and Mathematical Methods in Medicine . 2021;2021:1–12. doi: 10.1155/2021/5536903. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Qadri S. F., Zhao Z., Ai D., Ahmad M., Wang Y. Vertebrae segmentation via stacked sparse autoencoder from computed tomography images. Proceedings of the Eleventh International Conference on Digital Image Processing (ICDIP 2019); 2019; Guangzhou, China. [Google Scholar]
- 48.Ahmad M., Ding Y., Qadri S. F., Yang J. Convolutional-neural-network-based feature extraction for liver segmentation from CT images. Proceedings of the Eleventh International Conference on Digital Image Processing (ICDIP 2019); 2019; Guangzhou, China. [Google Scholar]
- 49.Qadri S. F., Shen L., Ahmad M., Qadri S., Zareen S. S., Khan S. OP-convNet: a patch classification-based framework for CT vertebrae segmentation. IEEE Access . 2021;9:158227–158240. doi: 10.1109/access.2021.3131216. [DOI] [Google Scholar]
- 50.Lei Y., Yuan W., Wang H., Wenhu Y., Bo W. A skin segmentation algorithm based on stacked autoencoders. IEEE Transactions on Multimedia . 2017;19(4):740–749. doi: 10.1109/tmm.2016.2638204. [DOI] [Google Scholar]
- 51.Krizhevsky A., Sutskever I., Hinton G. E. ImageNet classification with deep convolutional neural networks. Proceedings of the Advances in Neural Information Processing Systems; January 2012; Lake Tahoe, NV, USA. pp. 1097–1105. [Google Scholar]
- 52.LeCun Y., Bottou L., Bengio Y., Haffner P. Gradient-based learning applied to document recognition. Proceedings of the IEEE . 1998;86(11):2278–2324. doi: 10.1109/5.726791. [DOI] [Google Scholar]
- 53.Szegedy C., Toshev A., Erhan D. Deep neural networks for object detection. Proceedings of the Advances in Neural Information Processing Systems; 2013; Lake Tahoe, NV, USA. pp. 2553–2561. [Google Scholar]
- 54.Lu F., Wu F., Hu P., Peng Z., Kong D. Automatic 3D liver location and segmentation via convolutional neural network and graph cut. International Journal of Computer Assisted Radiology and Surgery . 2017;12(2):171–182. doi: 10.1007/s11548-016-1467-3. [DOI] [PubMed] [Google Scholar]
- 55.Wang B., Yang J., Ai J., et al. Accurate tumor segmentation via Octave convolution neural network. Frontiers in Medicine . 2021;8:p. 501. doi: 10.3389/fmed.2021.653913. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.He K., Liu X., Shahzad R., et al. Advanced deep learning approach to automatically segment malignant tumors and ablation zone in the liver with contrast-enhanced CT. Frontiers in Oncology . 2021;11:p. 2735. doi: 10.3389/fonc.2021.669437. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Ayalew Y. A., Fante K. A., Mohammed M. A. Modified U-Net for liver cancer segmentation from computed tomography images with a new class balancing method. BMC Biomedical Engineering . 2021;3(1):1–13. doi: 10.1186/s42490-021-00050-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Roth H. R., Lu L., Seff A., et al. A new 2.5 D representation for lymph node detection using random sets of deep convolutional neural network observations. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; June 2014; Boston, MA, USA. pp. 520–527. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Ciresan D., Giusti A., Gambardella L. M., Schmidhuber J. Deep neural networks segment neuronal membranes in electron microscopy images. Proceedings of the Advances in Neural Information Processing Systems; December 2012; New York, NY, USA. pp. 2843–2851. [Google Scholar]
- 60.Nasr M., Islam M. M., Shehata S., Karray F., Quintana Y. Smart healthcare in the age of AI: recent advances, challenges, and future prospects. IEEE Access . 2021;9 [Google Scholar]
- 61.Islam M. Z., Islam M. M., Asraf A. A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. Informatics in Medicine Unlocked . 2020;20 doi: 10.1016/j.imu.2020.100412.100412 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Muhammad L. J., Islam M. M., Usman S. S., Ayon S. I. Predictive data mining models for novel coronavirus (COVID-19) infected patients’ recovery. SN Computer Science . 2020;1(4):206–207. doi: 10.1007/s42979-020-00216-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Ayon S. I., Islam M. Diabetes prediction: a deep learning approach. International Journal of Information Engineering & Electronic Business . 2019;11(2):21–27. doi: 10.5815/ijieeb.2019.02.03. [DOI] [Google Scholar]
- 64.Asraf A., Islam M. Z., Haque M. R., Islam M. M. Deep learning applications to combat novel coronavirus (COVID-19) pandemic. SN Computer Science . 2020;1(6):363–367. doi: 10.1007/s42979-020-00383-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Saha P., Sadi M. S., Islam M. M. EMCNet: automated COVID-19 diagnosis from X-ray images using convolutional neural network and ensemble of machine learning classifiers. Informatics in Medicine Unlocked . 2021;22 doi: 10.1016/j.imu.2020.100505.100505 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Islam M. M., Iqbal H., Haque M. R., Hasan M. K. Prediction of breast cancer using support vector machine and K-nearest neighbors. Proceedings of the IEEE Region 10 Humanitarian Technology Conference (R10-HTC); December 2017; Dhaka, Bangladesh. pp. 226–229. [Google Scholar]
- 67.Haque M. R., Islam M. M., Iqbal H., Reza M. S., Hasan M. K. Performance evaluation of random forests and artificial neural networks for the classification of liver disorder. Proceedings of the International Conference on Computer, Communication, Chemical, Material and Electronic Engineering (IC4ME2); February 2018; Rajshahi, Bangladesh. pp. 1–5. [Google Scholar]
- 68.Ronneberger O., Fischer P., Brox T. U-net: convolutional networks for biomedical image segmentation. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; May 2015; Munich, Germany. pp. 234–241. [Google Scholar]
- 69.Kamnitsas K., Ledig C., Newcombe V. F. J., et al. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Medical Image Analysis . 2017;36:61–78. doi: 10.1016/j.media.2016.10.004. [DOI] [PubMed] [Google Scholar]
- 70.Zeiler M. D., Fergus R. Visualizing and understanding convolutional networks. Proceedings of the European Conference on Computer Vision; 2014; Zurich, Switzerland. pp. 818–833. [Google Scholar]
- 71.Candes E. J., Tao T. Near-optimal signal recovery from random projections: universal encoding strategies? IEEE Transactions on Information Theory . 2006;52(12):5406–5425. doi: 10.1109/tit.2006.885507. [DOI] [Google Scholar]
- 72.Plan Y., Vershynin R. Robust 1-bit compressed sensing and sparse logistic regression: a convex programming approach. IEEE Transactions on Information Theory . 2013;59(1):482–494. doi: 10.1109/tit.2012.2207945. [DOI] [Google Scholar]
- 73.Mendelson S., Pajor A., Tomczak-Jaegermann N. Uniform uncertainty principle for Bernoulli and subgaussian ensembles. Constructive Approximation . 2008;28(3):277–289. doi: 10.1007/s00365-007-9005-8. [DOI] [Google Scholar]
- 74.Yamaguchi O., Fukui K., Maeda K.-i. Face recognition using temporal image sequence. Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition; April 1998; IEEE, Nara, Japan. pp. 318–323. [Google Scholar]
- 75.Wolf L., Shashua A. Learning over sets using kernel principal angles. Journal of Machine Learning Research . 2003;4:913–931. [Google Scholar]
- 76.Elhamifar E., Vidal R. Sparse subspace clustering: algorithm, theory, and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence . 2013;35(11):2765–2781. doi: 10.1109/tpami.2013.57. [DOI] [PubMed] [Google Scholar]
- 77.Qiu Q., Sapiro G. Learning transformations for clustering and classification. The Journal of Machine Learning Research . 2015;16(1):187–225. [Google Scholar]
- 78.Giryes R., Sapiro G., Bronstein A. M. Deep neural networks with random Gaussian weights: a universal classification strategy? IEEE Transactions on Signal Processing . 2016;64(13):3444–3457. doi: 10.1109/tsp.2016.2546221. [DOI] [Google Scholar]
- 79.Amin J., Sharif M., Yasmin M., Fernandes S. L. Big data analysis for brain tumor detection: deep convolutional neural networks. Future Generation Computer Systems . 2018;87:290–297. doi: 10.1016/j.future.2018.04.065. [DOI] [Google Scholar]
- 80.Sutton R. Two problems with back propagation and other steepest descent learning procedures for networks. Proceedings of the Eighth Annual Conference of the Cognitive Science Society; 1986; Amherst, MA, USA. pp. 823–832. [Google Scholar]
- 81.Qian N. On the momentum term in gradient descent learning algorithms. Neural Networks . 1999;12(1):145–151. doi: 10.1016/s0893-6080(98)00116-6. [DOI] [PubMed] [Google Scholar]
- 82.Moghbel M., Mashohor S., Mahmud R., Saripan M. I. Automatic liver segmentation on Computed Tomography using random walkers for treatment planning. EXCLI Journal . 2016;15:500–517. doi: 10.17179/excli2016-473. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data supporting the findings of this study are included within the article.
