Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2021 Jan 18;2021:6235319. doi: 10.1155/2021/6235319

Sonar Objective Detection Based on Dilated Separable Densely Connected CNNs and Quantum-Behaved PSO Algorithm

Zhen Wang 1, Buhong Wang 1,, Jianxin Guo 2, Shanwen Zhang 2
PMCID: PMC7837782  PMID: 33531891

Abstract

Underwater sonar objective detection plays an important role in the field of ocean exploration. In order to solve the problem of sonar objective detection under the complex environment, a sonar objective detection method is proposed based on dilated separable densely connected convolutional neural networks (DS-CNNs) and quantum-behaved particle swarm optimization (QPSO) algorithm. Firstly, the dilated separable convolution kernel is proposed to extend the local receptive field and enhance the feature extraction ability of the convolution layers. Secondly, based on the linear interpolation algorithm, a multisampling pooling (MS-pooling) operation is proposed to reduce the feature information loss and restore image resolution. At last, with contraction-expansion factor and difference variance in the traditional particle swarm optimization algorithm introduced, the QPSO algorithm is employed to optimize the weight parameters of the network model. The proposed method is validated on the sonar image dataset and is compared with other existing methods. Using DS-CNNs to detect different kinds of sonar objectives, the experiments shows that the detection accuracy of DS-CNNs reaches 96.98% and DS-CNNs have better detection effect and stronger robustness.

1. Introduction

Ocean is an essential part of the Earth's life support system and is a huge natural treasure house with rich mineral and biological resources [1]. With the development of the world economy, the requirement in the field of national defense and national economy, such as marine resource development and dam safety detection, makes underwater imaging detection technology receive more and more attention [2]. Underwater imaging technology includes optical imaging and sonar imaging. The optical imaging effect is relatively close, and the imaging effect is unfavorable in the muddy water area [3]. Sonar imaging has the advantages of remote action and strong penetrating ability. It is especially suitable for muddy water area, so it is widely used in marine detection, underwater objective detection, and dam defect detection [4]. Due to the chaotic water medium of the underwater acoustic channel and the long transmission of the acoustic wave itself, the images acquired by the imaging sonar often have the problems of huge noise, severe distortion, blurred target edge, and low resolution, which seriously affect the underwater objective detection and identification [5]. Some sonar images are shown in Figure 1.

Figure 1.

Figure 1

Original sonar images.

Since the limitations of underwater sonar images, many optical image processing methods must be modified to apply to underwater acoustic image processing. Many researches have proposed a series of methods for the detection of sonar objective. The combination of feature extraction algorithm and classifiers has been widely used in sonar objective detection. For example, Lianantonakis and Petillot [6] proposed an unsupervised active contour model for sonar image segmentation. In this method, by extracting the texture features of the side-scan sonar image and extracting the weight parameters of different texture features through the curve evolution equation, the weight parameters were used to construct the active contour model. Elias et al. [7] proposed a sonar target detection method based on image texture features and independent correlation analysis. This method uses a traditional feature extraction method to extract multiple texture features of sonar images and then uses principal component analysis to perform dimensional reduction on texture features. The sonar objective is detected using independent correlation analysis. Markov is a statistical image segmentation algorithm with advantages of less model parameters and strong spatial constraints [8]. Mignotte et al. [9] used a new iterative condition estimation method to detect the objective region of the sonar image. The method establishes the Markov model by using the mixed distribution theorem and the marker parameter estimation and then uses the maximum likelihood estimation and the least square method to calculate the Markov model parameters. And the sonar image is segmented using the established Markov model. Cacciola et al. [10] used the finite element analysis (FEA) design and Misfit minimization methods to detect abnormal objective data. Collet et al. [11] proposed a hierarchical Markov random field model to solve the problem of sonar target detection. By combining maximum likelihood estimation and least squares method, they proposed a proportional casual multigrid algorithm, and the traditional Markov is improved as a hierarchical model for the label field of the pyramid structure, which is capable of accurately segmenting submarine shadows and seafloor reverberation noise. Yan et al. [12] proposed a sonar objective detection method based on extreme learning machine. This method assigns multiple extracted features by the convolutional neural network as central pixel to one of the categories and then, based on the central pixel, detects the objective area of the sonar image. The snake model uses some control points that form a certain shape as a template and through the elastic deformation of the template itself, matches the local features of the image to achieve harmony, and completes the energy function minimization to achieve image segmentation [13]. Angiulli et al. [14] used the second-order parabolic equations based on mathematical statistics to model and detect abnormal objective. Peng and Guo [15] proposed an improved snake model to detect the sonar objective. The snake model is defined as an energy minimizing spline which is influenced by external constraint force, and it can guide the image forces to pull toward feature. Compared with the traditional snake model, this improved snake model greedy algorithm can converge to the contours more quickly and more stably. In order to solve the problem of sonar target detection, Wang et al. [16] proposed a target detection algorithm based on the neutrosophic set and diffusion maps. Firstly, the neutrosophic set image of the sonar image is obtained, and then, the single gray threshold algorithm is used to extract the diffusion map of the sonar image; finally, the target region is detected using the improved target score equation. The combination of particle swarm optimization (PSO) algorithm and clustering algorithm has been widely used in the field of sonar image segmentation. In order to solve the problem of low efficiency and precision of existing sonar image segmentation methods, Zhao et al. [17] proposed a sonar image segmentation method based on PSO and neutrosophic set. The method first constructs image grayscale in NS domain. The symbiotic matrix expresses the fine texture features of the image, and it is based on the two-dimensional maximum entropy principle. And then, it is optimized by the particle swarm optimization algorithm to obtain the optimal two-dimensional segmentation threshold vector to achieve accurate segmentation of the objective region of the sonar image. Guo et al. [18] used a multibit particle system to encode the particles and proposed a sonar objective detection algorithm based on heuristic particle swarm optimization and fuzzy clustering. It calculates the particle swarm adaptive variance based on its own optimal particles and global optimal particles and updates the particle position in real time using a multibit quantum revolving gate. The output of the improved PSO algorithm is used to initialize the K-means clustering center. Based on the idea of fuzzy membership matrix in FCM and combined with the isolated spatial information features of noise, the objective region of sonar image is segmented. All the above methods use the traditional image processing method to detect the sonar objective. Although some effects are obtained, there are many problems with these methods, such as complex algorithm structure and poor robustness.

In recently years, deep learning has been widely used in the field of image processing. Since it can extract the deep semantic features of images and does not need to manually set the classifier, deep learning has many advantages in image segmentation, detection, and recognition [1921]. Many researchers have applied deep learning to sonar image processing to detect sonar objective. Song et al. [22] constructed deep convolutional neural networks (CNNs) to segment the sonar image. In the training process, the model uses multiple convolution kernels of the same size as the input image to extract features. Then, the Markov model is used to optimize the results obtained by CNNs, and the trained network model is used to segment different types of sonar images. Valdenegro [23] proposed a sonar objective detection algorithm based on a convolution neural network. This method detects the underwater small objective area by the constructed deep network model and proposes the regional suggestion network to generate a suggestion box for the objective region. Finally, the threshold optimization algorithm is used to optimize the parameters of the network model. The results show that the method can complete the detection of multiple small targets detection. Zhu et al. [24] first use the convolution neural network to extract the deep semantic features of the sonar image and then train the SVM classifier based on the manually labeled data. In the process of objective detection, the target region is detected using the sliding window with the kernel. The experiment results indicate that its performance is better than the objective detection algorithm of histogram of oriented gradients (HOGs) and local binary pattern (LBP). Based on the CNNs, Dzieciuch et al. [25] proposed a sonar objective detection method based on a combination of nonlinear classifier and CNNs. This method can extract abstract and complex features of sonar images, and it can accurately detect the objective region with complex background environments. Since the method introduces nonlinear operations, it has a faster training speed and higher detection accuracy. As there are few sonar images that can be used for training, Chen and Summers [26] proposed an unsupervised CNNs model for sonar objective detection. The method constructs a semisupervised ladder network and uses it for a small sample sonar image to execute classification and positioning. The semisupervised semantic segmentation and objective localization of sonar images are realized by pixel-level classification of the entire sonar image and by using the ladder network to realize semantic segmentation and objective localization of sonar images. Since noise affects the segmentation accuracy of sonar images, Jiao et al. [27] proposed a sonar image segmentation method based on improved fully convolution neural networks (FCNs). This method improves the original mean square error loss function by using cosine distance. The improved method can complete the accurate segmentation of the submarine sonar image. Williams and Dugelay [28] proposed a method for multiview sonar image classification based on image fusion and deep learning methods. The method can accurately classify different types of sonar images and complete the classification task for sonar images containing noise and complex background. Wang and Zhang [29] proposed a method based on multiscale multicolumn convolution neural network (MSMC-CNNs) for sonar objective detection. The parallel network structure design of MSMC-CNNs can enhance the multiscale feature extraction ability, and experimental result shows it can complete the detection task for different sonar objectives. Bin et al. [30] proposed a sonar objective detection method based on transfer learning and geometric feature constraints. The method first uses transfer learning to improve the detection accuracy of the network model and then employs the geometric feature constraint method to improve the detection efficiency of the model. Finally, different kinds of sonar images are detected with the constructed model.

The evolutionary computing (EC) and swarm intelligence (SI) are mature global optimization methods with high robustness and wide applicability, which have the characteristics of self-organization, self-adaptation, and self-learning [31]. Moreover, the EC and SI algorithms are not limited by the nature of the problem and can effectively deal with complex optimization problems that are difficult to solve by traditional optimization algorithm. Based on the advantages of EC and SI, many researches used it to solve the weight parameter optimization problem of CNNs. Alejandro et al. [32] proposed the EvoDeep evolutionary algorithm to optimize the structure and weight parameters of CNNs, which can improve the classification accuracy and maintain a valid layers sequence. Watanabe and Shinozaki [33] used genetic algorithm (GA) and CMA-ES algorithm to optimize the weight parameters of CNNs. In this method, the GA is used to represent CNN structure as directed acyclic graphs and optimized the graph structure by CMA-ES algorithm. Real et al. [34] proposed a new neural network evolution method to evolve the structure of CNNs. This method can effectively optimize the weight parameters of CNNs, which achieve high classification accuracy on the CIFAR-10 and CIFAR-100 datasets. Chung and Shin [35] proposed a GA-optimized multichannel CNN method for stock market prediction. The method uses the GA algorithm to optimize the CNN parameters, and the experimental results show that GA algorithm can effectively improve the compute efficiency and prediction accuracy. To solve the problem of CNN optimization, Połap et al. [36] used a multithreading mechanism for weight parameters selection, in which the multicore mechanism is used to select the best weight among all parallel training networks. Moayedi et al. [37] optimized the artificial neural network (ANN) by the combination of dragonfly algorithm (DA), whale optimization algorithm (WOA), and invasive weed optimization (IWO). Experimental results show that this method can effectively improve the prediction accuracy of the ANN model. The above methods show that the use of EC and SI algorithms can effectively optimize the structure and weight parameters of CNNs.

At present, the sonar objective detection method based on deep learning has better effect than the traditional image processing method. However, due to the complicated structure of the model and the long training time of the weight parameter, the existing models are less practical. In order to solve the problems of deep learning method in sonar objective detection, this paper is based on U-Net, dilated convolution, separable convolution, and DenseNet to propose dilated separable densely connected convolution neural networks (DS-CNNs) for sonar objective detection. The main contributions of this paper are listed as follows:

  1. The dilated separable convolution kernel is proposed to expand the local receptive field of the traditional convolution kernel and enhance the feature extraction ability of the convolution layers.

  2. Based on the linear interpolation algorithm, a multisampling pooling method is proposed to solve the problem of feature and sample information loss caused by convolution operation.

  3. The quantum-behaved particle swarm optimization (QPSO) algorithm is proposed to replace the traditional backpropagation (BP) algorithm to train the weight parameters of the network model.

  4. The DS-CNN is validated and compared with the state-of-the-art methods on the sonar image dataset, and the robustness of DS-CNNs is analyzed.

The rest of this paper is organized as follows. In Section 2, the related work is introduced, including U-Net, dilated convolution, separable convolutions, and DenseNet. In Section 3, DS-CNNs are described in detail, including model parameters, multisampling pooling, and QPSO algorithm. The experiments on the sonar image dataset are implemented in Section 4, and the conclusion is given in Section 5.

2. Related Works

In this section, the related works are introduced, which includes U-Net, separable convolution, dilated convolution, and DenseNet.

2.1. U-Net Network Model

The structure of U-Net [38] is the same as that of FCN [39] and SegNet [40], which includes two parts of mutually symmetric contraction network and expansion network. The contraction network part in U-Net includes convolutional layers, pooling layers, and activation function, whose main function is to complete feature extraction of the input image. The expanded network portion includes deconvolution and upsampling and is mainly used to restore the size resolution of the feature map. Moreover, the uses of skip connections and pooling indexes in the U-Net structure can better improve the segmentation accuracy. The U-Net structure is shown in Figure 2. It can be seen from Figure 2 that the U-Net contraction network and the expansion network form a U-shaped structure. The convolution and pooling operations of the contraction network can extract the multilayer semantic information of the image, while the expansion network realizes the restoration of the image features through multilayer deconvolution and upsampling operations, and the expanded network is restored by the pooled index operation, so the boundary information of the feature map is continuously restored during the process.

Figure 2.

Figure 2

Structure of U-Net network model.

U-Net is characterized by the mutual mapping of the contraction network and the expansion network. In the expansion network, the image feature map in the contraction network is merged and the missing boundary information is complemented, thereby improving the classification accuracy of the pixels. Moreover, U-Net's U-shaped structure makes the model more intuitive for the cutting and splicing process of feature maps. By splicing high-level semantic features with low-level semantic features, the model can obtain more context information and detailed information. The contraction network in U-Net consists of ten convolutional layers and four max-pooling layers. Each two convolutional layers and one max-pooling layer form one downsampling block, and there are five downsampling block structures in the contraction network. Each downsampling block is composed of two 3 × 3 sized convolution layers and one 2 × 2 sized max-pooling layer. After each downsampling block operation, the size of the output feature map turns out to be one-half of the input feature images, and the number of channels becomes twice of the original. The extension network and the contraction network correspond to each other, and they are mainly used for image upsampling operations. They include five upsampling blocks of the same structure, each of which contains two 3 × 3 convolution layers and one 2 × 2 max-pooling layer. The deconvolution layer uses a nonlinear unit as the activation function. After each upsampling, the image becomes twice as large as the original feature maps, and the number of channels becomes one-half. Each pixel is classified at the end of the network using SoftMax classifier.

2.2. Depth-Wise Separable Convolutions

Traditional convolutional neural networks use large-scale convolution kernels. For example, AlexNet uses an 11 × 11 size convolution kernel. Although a large local receptive field is obtained, the calculation of the parameters in the model also increases. In the GoogLeNet [41] model, multiple 3 × 3 small-scale convolution kernel cascade structures are used to reduce the parameter calculation of the model under the same conditions to maintain the original image receptive field. The small-scale convolution kernel cascade structure not only increases the depth of the network model but also enhances the nonlinear expression ability of image features. Since the traditional convolution kernel needs to convolute each channel of the input feature maps, the parameter calculation of the model is heavy. In order to solve the problem of large amount of parameter calculation, Howard et al. [42] proposed a depth-wise separable convolution kernel, which can greatly reduce the parameter calculation of the network model. The depth-wise separable convolution firstly uses two-dimensional convolution operation on each channel in the input feature map to reduce the amount of parameter calculation and then uses 1 × 1 convolution kernel to carry out three-dimensional convolution operations for the combination of different channels. The structure of the depth-wise separable convolution is shown in Figure 3. Assume that the size of the input feature maps is Df × Df, the number of channels is M, and the size of the convolution kernel is DM × DM, which contain N convolution kernels.

Figure 3.

Figure 3

Depth-wise separable convolutions kernels.

2.3. Dilated Convolutions

The use of standard convolution kernels in convolutional layers can lead to insufficient information extraction capabilities in the process of feature extraction of images. Moreover, when the feature reduction is performed using the pooling operation, although the parameter calculation amount of the network model is reduced, a large amount of useful information also get lost. The use of dilated convolution can effectively avoid the information loss caused by the pooling operation, and it can also expand the local receptive field of the convolution kernel [43]. Moreover, the dilated convolution does not need to compress the resolution of the image and retains a large amount of internal space information in the image, which has better effect in image segmentation than conventional convolution kernel. The dilated convolution contains the dilated coefficient to control the size. The larger the dilated coefficient is, the larger the receptive field corresponding to the convolution kernel of the same calculation amount is. As shown in Figure 4, the dilated convolution kernels of different dilated coefficients are, respectively, indicated, where the red dot area indicates the position where the convolution operation is required, and the blue area indicates the local receptive field of the dilated convolution. It can be seen from Figure 4 that the dilated convolution kernel gradually increases with the increase in the dilated coefficient.

Figure 4.

Figure 4

Dilated convolution kernel structures with three dilation rate parameters: (a) coefficient is 1; (b) coefficient is 2; (c) coefficient is 3.

2.4. DenseNet

In order to enhance the feature extraction ability of convolutional neural networks, Huang et al. [44] proposed the densely connected convolutional networks (DenseNet) structure based on ResNet [45]. DenseNet uses a simpler feature transfer method to achieve feature reuse by connecting shallow features to deep features. The DenseNet network structure is shown in Figure 5. The output of each layer can be directly connected to the input of each layer so that the method reduces the parameters of the network model while increasing the network depth. Using densely connected structure can reduce the number of parameters, but pooling operations in the network produce feature maps of different sizes, making it impossible to connect them directly. In order to solve the problem that the feature layers of the overall structure cannot be densely connected, DenseNet employs a block design approach, using dense connections only in each block area. The output of each block will be used as the input to the next block after the pooling operation, and the densely connected operation will still be used in the next block. At the same time, a 1 × 1 convolutional layer is added to each block to reduce the thickness of the feature maps and reduce the parameter size of the network model. The overall structure of DenseNet is different from the traditional networks model. Each block of DenseNet contains a batch normalization (BN), ReLU activation function, and convolution layer. Moreover, an additional convolutional layer is added at the beginning of the DenseNet structure. This operation enables the BN and the ReLU in each block to better function on the output of the previous block. Since DenseNet uses a method that directly passes features, it is easier for the network to preserve shallow features, which helps to produce accurate segmentation results.

Figure 5.

Figure 5

Densely connected structure.

3. Proposed Method

In this section, the proposed dilated separable densely connected convolutional neural network (DS-CNN) structure, quantum-behaved particle swarm optimization (QPSO) algorithm, and multisampling pooling (MS-pooling) are introduced.

3.1. DS-CNN Model

Based on U-Net [38], dilated convolution [41], depth-wise separable convolution [42], and DenseNet [44], a dilated separable densely connected convolutional neural networks (DS-CNNs) for sonar objective detection is proposed. DS-CNNs not only improve the detection accuracy of sonar objective but also reduce training and segmentation time. The structure of DS-CNNs is shown in Figure 6. DS-CNNs include encoder network structure and decoder network structure, where the input of the encoder network is the original sonar image and the encoder network is composed of five dilated separable convolution blocks. Each convolution block contains two dilated separable convolution layers. Each dilated separable convolution layer consists of depth-wise separable convolution of 3 × 3 and 1 × 1 sizes, with different dilated coefficients set in these depth-wise separable convolutions. Convolution kernel with a dilated coefficient of 1 is used to extract the detailed features of the image, and the convolution kernel with a dilated coefficient of 2 can extract the edge features of the image. The convolution combination of different dilated coefficients can capture the multiscale context information of the image. In order to avoid the gradient explosion problem caused by the increase in network depth and also to make better use of the feature information between different layers, the input and output of the dilated separable convolution blocks are connected by using densely connection and 1 × 1 convolution kernel. Moreover, the different dilated separable convolutional blocks in the encoder network are connected using the MS-pooling layers, and the MS-pooling size is 2 × 2. As the number of network layers increases, the encoder network's ability to characterize images also becomes stronger. Each time the convolution kernel is pooling, the channel number of the convolution kernel is doubled. The specific parameters are shown in Table 1.

Figure 6.

Figure 6

The architecture of DS-CNNs.

Table 1.

Specific parameters of DS-CNNs.

Layer Type Kernel Channel
Layer_1 Conv 3 × 3 32
Layer_2 Conv 3 × 3 32
Layer_3 Conv 1 × 1 32
Layer_4 MS-pooling 2 × 2 64
Layer_5 Conv 3 × 3 64
Layer_6 Conv 3 × 3 64
Layer_7 Conv 1 × 1 64
Layer_8 MS-pooling 2 × 2 128
Layer_9 Conv 3 × 3 128
Layer_10 Conv 3 × 3 128
Layer_11 Conv 1 × 1 128
Layer_12 MS-pooling 2 × 2 256
Layer_13 Conv 3 × 3 256
Layer_14 Conv 3 × 3 256
Layer_15 Conv 1 × 1 256
Layer_16 MS-pooling 2 × 2 512
Layer_17 Conv 3 × 3 512
Layer_18 Conv 3 × 3 512
Layer_19 Conv 1 × 1 512
Layer_20 MS-pooling 2 × 2 1024
Layer_21 DC
Layer_22 Conv 3 × 3 512
Layer_23 Conv 3 × 3 512
Layer_24 MS-pooling 2 × 2 512
Layer_25 Upsampling 2 × 2 256
Layer_26 DC
Layer_27 Conv 3 × 3 256
Layer_28 MS-pooling 2 × 2 256
Layer_29 Upsampling 2 × 2 128
Layer_30 DC
Layer_31 Conv 3 × 3 128
Layer_32 MS-pooling 2 × 2 128
Layer_33 Upsampling 2 × 2 64
Layer_34 DC
Layer_35 Conv 3 × 3 64
Layer_36 MS-pooling 2 × 2 64
Layer_37 Upsampling 2 × 2 32
Layer_38 DC
Layer_39 Conv 1 × 1 32
Layer_40 SoftMax 2 2

The decoder network structure performs upsampling on the feature map extracted in the encoder network and hierarchically maps the high-level semantic features extracted in the encoder network to a semantic segmentation map of the same size. The decoder network structure includes three dilated separable convolution blocks, and different convolution blocks are connected by bilinear interpolation sampling layer operation. In the upsampling process, the traditional fully convolutional neural networks only upsample the final mapping result to obtain an image of the same size as the input. However, this upsampling operation ignores the shallow output information, and this information also has some characterization capabilities for the image. Therefore, in order to better keep the spatial information of the image, DS-CNNs fuse the shallow layers information with the corresponding decoder network structure through densely connection (DC) so that the shallow context information can be propagated to a higher resolution. And the deep abstract features are combined with shallow representation features to locate image grayscale category information.

3.2. Multisampling Pooling

The purpose of the pooling operation in the convolutional neural network is to reduce the dimensionality of the convolutional feature map and selectively compress the feature map to further extract the convolutional features. The pooling options in CNNs include mean-pooling and max-pooling, wherein the mean-pooling is to obtain the average value of the convolution feature maps, and the max-pooling is to obtain the maximum value of the convolution feature maps. The calculation formula for mean-pooling and max-pooling is as follows:

out_putk=1m×mi=1mj=1min_putijk, (1)

where m represents the size of the pooling area, in_putijk represents the k − th pooling area, and out_putk represents the pooling feature value corresponding to the k − th pooling area.

out_putk=maxi,j1,min_putijk, (2)

where max(·) represents the maximum function, and the returned result is the maximum value in the in_putijk region.

In the process of using the mean-pooling and max-pooling, the mean-pooling is to use the mean value in the pooling area as the characteristic response, which is easy to cause the deviation of the estimated mean value in the feature map, resulting in the blurring effect. The max-pooling is to obtain the characteristic response by obtaining the maximum value in the pooling area, so it is easy to ignore the detailed features extracted in each pooling layer. In order to solve the problems of the traditional pooling operation, multisampling pooling (MS-pooling) method is proposed instead of the original pooling operation. MS-pooling is proposed based on the interpolation algorithm, so it is necessary to insert a basis function to fit the data in MS-pooling. The corresponding value of the function f(·) at the pixel point p(x, y) is obtained by weighted average of the sixteen points closest thereto, and MS-pooling uses two polynomial interpolation cubic functions of the sampling points in the rectangular grid, among which the basic functions are as follows:

Sw=12w2+w3,w<1,48w+5w2w3,1w<2,0,w2. (3)

The calculation method of multisampling pooling is as follows:

fi+u,j+v=ABC, (4)

where A, B, and C all represent matrices and the specific form is as follows:

A=S1+uSuS1uS2u,B=fi1,j2fi,j1fi+1,j2fi+2,j2fi1,j1fi,j1fi+1,j1fi+2,j1fi1,jfi,jfi+1,jfi+2,jfi1,j+1fi,j1fi+1,j1fi+2,j+1,C=S1+vSvS1vS2v, (5)

where f(i, j) represents the pixel value at the pixel point t(i, j), f(i+u, j+v) represents the pixel coordinates in the original image, and u and v represent the pixel coordinates in the both horizontal and vertical directions.

In the process of using MS-pooling operations, assume that the original image size is m × m and the image size after using MS-pooling is M × M. According to the ratio between the original image and MS-pooling image, the final coordinates of the image F(X, Y) are determined by the following formula:

Ox,y=OXMm,YMm=fi+u,j+v. (6)

Convert it into a MS-pooling function, whose function form is as follows:

F=interpolation_poolO,M,M, (7)

where interpolation_pool represents the MS-pooling operation, O is the input of the MS-pooling operation, F is the output of the MS-pooling operation, and [M, M] represents the image size using the MS-pooling operation.

The pooling effect of different pooling operations is shown in Figure 7. The original image is first subjected to convolution operation to obtain a convolutional feature map, and then, a pooling feature map is obtained through pooling operation. Compared with the mean-pooling and the max-pooling, the MS-pooling maintains its original features as much as possible and can effectively avoid the transmission of its features in the multiple layer networks without causing feature loss and gradient disappearance.

Figure 7.

Figure 7

Different pooling operation results: (a) original image; (b) convolution feature; (c) mean-pooling; (d) max-pooling; (e) MS-pooling.

3.3. QPSO Optimization Algorithm

The traditional CNNs model uses the backpropagation (BP) algorithm to complete weight parameter optimization. The BP algorithm tends to make the CNNs model training fall into local optimum, and the convergence speed is slow, which cannot guarantee the generalization ability of the network model. To solve the problem of BP algorithm, we propose the quantum-behaved particle swarm optimization (QPSO) algorithm to train the CNNs model. In the traditional particle swarm optimization (PSO) algorithm, the problem to be solved in optimization is to define a particle point in the N-dimensional search space. The particles fly in the search space at a certain speed and dynamically adjust the flight process according to their flight experience. In the process of flight search, each particle uses its current position and current speed, the distance between current position and particle best, and the distance between current position and global best to change its particle position. In the standard PSO algorithm, the particle swarm consists of m particles, and the position of each particle is assumed to be a possible preliminary solution to the problem in the N-dimensional search space. Particles update their flight trajectory based on their inertia, optimal position, and optimal group position.

The idea of the PSO algorithm is to accelerate the close proximity of each particle to itself and the group. The moving process is shown in Figure 8. The starting position and velocity of the particle are randomly set in the solution space. During the iterative process, the PSO algorithm will record the optimal position experienced by individual particles and groups and the corresponding fitness function values. There are many problems in the traditional PSO algorithm, such as slow convergence and easiness to fall into local optimum and poor processing for discrete optimization problems. To solve the defects of PSO algorithm, we propose quantum-behaved particle swarm optimization (QPSO) algorithm. Based on the traditional PSO algorithm, we introduce the contraction-expansion factor and the differential evolution operator in the QPSO algorithm, which not only accelerates the optimization speed of the PSO algorithm but also prevents the algorithm from falling into the local optimal solution. The calculation formulas of the QPSO algorithm are as follows:

pid=θ·pbestid+1θ·gbestd, (8)

where pid represents the random position of particle i between pbestid and gbestdA.

mbest=1Mi=1Mpbesti, (9)

where mbest represents the center point of the current best position for all individuals.

kt=PgtavgPit, (10)

where Pg(t) represents the optimal position of the population and avgPi(t) represents the optimal average position of the individual.

αt=αoαm·exptT2·kt+αm, (11)

where αo represents the initial value of the contraction-expansion factor, αm represents the final value of the contraction-expansion factor, t represents the current number of iterations, and T represents the maximum number of iterations.

Xid=θ·δd+1θ·gbestd±α·mbestdxid·ln1u, (12)
δ=XKXj,ijk, (13)

where Xid represents the position of the next generation particle i, δ represents the evolution operator, gbestd represents the global optimum of all particle positions in the current generation, xid represents the position of the current generation of the particle i, and θ and u represent the uniform distribution in the 0 to 1 between. Training the DS-CNNs using the QPSO algorithm requires calculating the position information of the particles according to equations (8) to (13) and updating the weights and thresholds of the network.

Figure 8.

Figure 8

Diagram of particles moving in the PSO algorithm.

The training process of DS-CNNs based on the QPSO algorithm is as follows:

  • (1)

    Design the structure of the network model and initialize the weight parameters of the DS-CNNs.

  • (2)

    Initialize the position information of the particle swarm.

  • (3)

    Set the control shrink-expansion factor and initialize the NU ∈ (0,1).

  • (4)

    The average optimal position of the particle swarm is calculated according to equation (9).

  • (5)

    Randomly take values from 0 to 1 and determine if it is less than NU and update the particle position according to equations (11) and (12).

  • (6)

    The particles are mapped to the weight parameters of each layer according to the principle of mutual correspondence.

  • (7)
    Calculate the mean square error of the network and sort the particles according to their fitness. The mean square error function is calculated as follows:
    Net.E=1Ni=1Nj=1HYjiDyji2, (14)
  • where N represents the number of training samples input, Hrepresents the number of neurons in the output layer, YjiD represents the expected output of the j output node of the first i sample, and yji represents the actual network output of the j node of the first sample.

  • (8)

    Update particle individual optimality and global optimality.

  • (9)

    Confirm whether the termination condition is met. If it is satisfied, exit the loop and output the global optimum, and use it as the weight parameter of the network. Otherwise, loop again to find the optimal solution.

3.4. Evaluation Indicator

In order to quantitatively analyze the segmentation accuracy of sonar images, the specificity (SP), sensitivity (SE), accuracy (Acc), Jaccard index (Jacc), and dice coefficient (dice) are used as evaluation criteria. SP represents the proportion of the negative values considered by the segmentation method and the values that are really negative in the ground truth. SE represents the proportion of the positive values considered by the segmentation method and the real positives values given by the ground truth. Acc represents how close the ground truth and the segmentation proposed by the methods are and the relation between their hits and errors. Jacc is a statistic metric that measures the similarity between the ground truth and the segmentation proposed by the methods. Dice is used to measure the similarity between two samples. The calculation formulas are as follows:

SP=NtnNtn+Nfp,SE=NtpNtp+Nfn,Acc=Ntp+NtnNtp+Nfp+Nfn+Ntn,Jacc=NtpNtp+Nfn+Nfp,Dice=2·Ntp2·Ntp+Nfn+Nfp, (15)

where Ntp represents the number of true positive pixels and a true positive indicates that the pixel belongs to the objective area both in the predicted segmentation result and in the ground-truth segmentation result. The Nfp represents the number of false positive pixels, and a false positive indicates that the pixel is an objective area in ground-truth, but it is a background area in predicting the segmentation result. The Ntn represents the number of true negative pixel points. True negative indicates that the pixel is a background area in the ground-truth, and the predicted result is also a background area. The Nfn represents the number of false positive pixels, and a false positive indicates that the pixel is a background region in the ground-truth, but it is an objective region in the predicted segmentation result.

4. Experiment and Analysis

All experiments are conducted with Intel Xeon E5-2643v3@3.40 GHz CPU, 64 GB RAM, NVidia Quadro M4000 GPU, 8 GB of video memory, by CUDA Toolkit 9.0, CUDNN V7.0, Python 3.5.2, Tensorflow-GPU 1.8.0, Windows 7 64 bit operating system.

4.1. Benchmark Datasets

Sonar image dataset is collected by dual-frequency side-scan sonar (Shark-S450D), which is produced by Canadian Imagenex Company. The collection location is in Qingdao, China. Some original sonar images are shown in Figure 9. To use the sonar images dataset to train the weight parameters of the network model, the collected 3600 sonar images are proportionally divided into train dataset, verification dataset, and test dataset. We randomly select 60% of the dataset as training dataset, 30% of the sonar image dataset as verification dataset, and the other 10% as test dataset. Specifically, the train dataset includes 2160 sonar images, the verification dataset includes 1080 images, and the test dataset includes 360 images.

Figure 9.

Figure 9

Original sonar image.

The average image size is 12000 × 8200 pixels, where the pixel values in the image are determined by the reflected strength of the echo and the distance from objects. Since the original sonar image contains much noise and is affected by the seabed environment and the sonar equipment, the collected images quality is poor. Therefore, the sonar image needs to be preprocessed. Firstly, the Gaussian filtering algorithm is used to denoise the image. Specifically, the sonar image is filtered using a 3 × 3 Gaussian filter template. Secondly, the linear interpolation algorithm is used to enhance the objective area in the sonar image. The preprocessed sonar images are shown in Figure 10.

Figure 10.

Figure 10

Sonar image preprocessing: (a) original image; (b) denoise; (c) enhance.

Since the original sonar images size is too large, it will occupy more GPU resources and increase the training time of the network models, and the original sonar image is reduced to 500 × 500 pixels. But directly reducing the size of the sonar image will result in loss of the information of the image. So we use a patch-wise method to solve this problem. As the ground truth for images is essential to use the supervised method for sonar image detection, we use the image annotation tool LabelMe, which is developed by the MIT to manually label images. The project source code is open. The original image and the corresponding manually annotated image are shown in Figure 11, where red represents the objective area and black represents the background area.

Figure 11.

Figure 11

Original sonar images and corresponding labeled images: (a) original image; (b) manually label; (c) ground truth.

4.2. Network Model Training

In the training process of the network model, the BP algorithm is usually used to optimize the weight parameters of the network model, but in the DS-CNNs, the QPSO algorithm is used to optimize the weight parameters. In order to verify the superiority of QPSO algorithm in DS-CNN training, the BP algorithm and QPSO algorithm are used to train the network model in the sonar dataset. In the process of training, the batch size is set to 64, the number of epoch is 3000, and the accuracy is used as the segmentation evaluation standard. The training process and detection accuracy are shown in Figure 12.

Figure 12.

Figure 12

Training effect of different optimization algorithms.

It can be seen from Figure 12 that QPSO algorithm converges faster than BP algorithm in the training process. When the whole training is completed, the loss value of the QPSO algorithm is smaller than that of BP algorithm, indicating that the QPSO algorithm is suitable for training DS-CNNs. In the comparison of the segmentation accuracy results, the accuracy of QPSO algorithm is always higher than BP algorithm in the training process.

Due to the small size of the sonar dataset, the direct use of the network model training will lead to overfitting or low segmentation accuracy. The transfer learning is to transplant the model parameters trained in the large dataset to train ImageNet. In the network model, the training efficiency and robustness of the model can be improved. Therefore, in the training process of the DS-CNN model, the transfer learning method is introduced, and the network model parameters of the original U-Net network after repeated training in the large image dataset are used for the detection of sonar images. Common transfer learning methods include feature transfer and parameter transfer. Feature transfer is to eliminate the classification layer of the pretraining network and transform the feature map output from the convolution layer and the pooling layer into feature vector. And then, it sends feature vector to a new classifier for detection training. The parameter transfer is to randomly initialize the partial convolution layer and the remaining convolutional layers directly using the weight parameters of the pretrained network model and use the dataset to be trained to retrain the network parameters. In DS-CNN model, it is trained using a parameter transfer method for sonar detection. The initial learning rate of the network model is set to 0.001, the regular term coefficient is set to 0.001, and the number of epoch is set to 3000. The training effect of the two training methods of the new learning and transfer learning is compared in the sonar images dataset. The effect of the different training methods is shown in Figure 13.

Figure 13.

Figure 13

Comparison of different learning methods.

It can be seen from Figure 13 that the use of transfer learning accelerates the convergence speed of the network model. As the transfer learning enhances the feature extraction capability, the network model achieves a higher detection accuracy.

4.3. Comparison of Different Convolution Kernels

In order to verify that the proposed dilated separable convolution kernel has a stronger feature extraction ability in sonar image and it is more suitable for sonar objective detective tasks, the traditional convolution, dilated convolution, deformable convolution, and dilated separable are used to compare the feature extraction effects of the sonar image. The feature extraction visualization results are shown in Figure 14. Since different convolution kernels have different regions of interest (ROIs) in the image, the heat map is used to visualize the convolution kernel feature extraction region. The heat map of the different convolution kernels is shown in Figure 15.

Figure 14.

Figure 14

Feature extraction effect of different convolution kernels: (a) original image; (b) traditional; (c) dilated; (d) deformable; (e) dilated separable.

Figure 15.

Figure 15

ROI of different convolution kernels: (a) original image; (b) traditional; (c) dilated; (d) deformable; (e) dilated separable.

It can be seen from Figure 14 the traditional convolution kernel has a poor feature extraction effect on the sonar image and a large amount of detailed image features are lost, resulting in an inconspicuous outline of the objective area. Since the dilated convolution uses different dilated coefficients to enlarge the local receptive field of the convolution kernel and enhance the feature extraction ability of the convolution kernel, it can extract the contour boundary of the objective, but the extracted boundary is fuzzy, and many related information is lost. Deformable convolution is poorly effective in feature extraction of sonar images. Since it is affected by noise in the sonar image, the deformable convolutions cannot completely extract feature for contour boundary feature extraction and detail feature extraction. The dilated separable convolution kernel not only extracts the contour features of the sonar image objective area but also enhances the ability to extract the detailed features by using the separable convolution, so it is suitable for the feature extraction task of the sonar images. It can be seen from Figure 15 that the ROI of the dilated separable convolution is concentrated, indicating that the dilated separable convolution has a better feature extraction effect on the objective area of the sonar image. Since different convolution kernels have an impact on the detection accuracy of sonar images, in order to verify the effectiveness of the dilated separable convolution for sonar image detection, we use traditional convolution kernel, dilated convolution kernel, deformable convolution kernel, and dilated separable convolution kernel to perform detection experiment in a sonar image dataset. The accuracy of detection is shown in Figure 16.

Figure 16.

Figure 16

Detection accuracy of different convolution kernels.

It can be seen from Figure 16 that the dilated separable convolution has a higher detection accuracy, and because it uses a separable structure, the detection accuracy in the first 500 epoch is higher than that of other convolution kernels. After 3000 epochs are completed, the final detection accuracy reaches 95%.

4.4. Comparison of Different Detection Methods

In order to verify the effectiveness of DS-CNNs in sonar objective detection, DS-CNNs were compared with existing U-Net [38], FCN [39], SegNet [40], DeepLabV3 [43], and PSPNet [46] network models on sonar image datasets. The training process and detection accuracy of different methods are shown in Figure 17. The sonar objective detection experiment was performed on the sonar image dataset using DS-CNNs and other detection methods, and some of the visual results are shown in Figure 18.

Figure 17.

Figure 17

Training process and detection accuracy of different methods: (a) training process; (b) detection accuracy.

Figure 18.

Figure 18

Sonar images detection results of different pixel sizes: (a) original; (b) FCN; (c) SegNet; (d) U-Net; (e) DeepLab; (f) PSPNet; (g) DS-CNNs.

It can be seen from Figure 17(a) that, during the training process, the DS-CNNs converge faster and the loss value is less fluctuating. FCN and SegNet converge slowly during training, and their loss value is still high when training is completed. U-Net converges faster during the initial training, but when it continues to train, the loss value gradually increases. The training process of DeepLabV3 and PSPNet is relatively stable, but the loss value is higher when the train is completed, which indicates that the training effect is poor under the same epoch times. The reason why the training effect is better is that it employs the dilated convolution kernel, which can reduce the parameter calculation of the network model and accelerates the convergence speed in the training process. Figure 17(b) shows that DS-CNNs have a higher accuracy than the other methods in the detection of sonar objectives, indicating that DS-CNNs are suitable for sonar objectives detection tasks.

Figure 18 shows the results of different models on the sonar image test dataset. From the test results that the FCN, SegNet, and U-Net models can detect the contour region of the sonar objectives, the detection effect on the detailed information is poor, and there is information loss of the objective region. The DeepLab and PSPNet models have better segmentation of the detail feature of the objective region, but the effect is not good when detecting the background area. The results in Table 2 show that DS-CNNs have better effect in the sonar objective detection task in different evaluation indicators, and the average detection accuracy reaches 96.98%, indicating that DS-CNNs are suitable for sonar objective detection.

Table 2.

Detection accuracy of different models.

Model SP (%) SE (%) Acc (%) Jacc (%) Dice (%)
FCN 78.61 81.38 89.62 87.32 85.86
SegNet 79.26 83.25 91.02 88.97 87.06
U-Net 81.38 85.93 92.28 90.03 88.65
DeepLab 80.26 84.38 91.92 89.37 87.84
PSPNet 83.39 87.65 94.18 92.25 90.06
DS-CNNs 85.89 89.37 96.98 94.48 92.89

The calculation efficiency of different models is evaluated using the evaluation index of memory space, the number of parameters, training times, and test time. The calculation performance of different models is shown in Table 3. As can be seen from Table 3, DS-CNNs have the minimum values for different calculation efficiency evaluation indicators. The memory space during the training process is 3.42 GB, which ensures that DS-CNNs can be deployed on multiple mobile platforms. Moreover, the training time of DS-CNNs is 4.5 h, which indicates that the model is practical. The purpose of proposed DS-CNNs is to use it for underwater target detection. In the actual application process, underwater detection equipment must detect the location and contour of the underwater target timely and accurately, so the higher requirements are put forward for the real-time performance of the detection algorithm. The single image test time of DS-CNNs is 1.5 s, which can meet the real-time requirements of underwater objective detection.

Table 3.

Calculation efficiency of different models.

Model Memory space (GB) Training time (h) Parameters
FCN 8.45 12.6 63,254,618
SegNet 6.13 10.3 56,421,514
U-Net 5.42 8.6 48,163,254
DeepLab 7.13 7.3 52,618,315
PSPNet 4.26 7.8 45,126,257
DS-CNNs 3.42 4.5 35,183,326

4.5. Robust Analysis

Since the sonar image is taken underwater, the underwater complex environment will have an impact on the quality of the sonar image. Similarly, the quality of the sonar image also has an impact on the experiment results. In order to verify the effectiveness of the DS-CNNs, the proposed method is used to detect the sonar images of different pixel sizes (Ps), and the experiment results are shown in Figures 19 and 20. It can be seen from that as the pixel size increases, the detection accuracy of DS-CNNs increases gradually, which indicates that the pixel size has an impact on the detection accuracy, but the detection accuracy of DS-CNNs is around 95%, indicating that DS-CNNs are less affected by changes in pixel size.

Figure 19.

Figure 19

Sonar images detection results of different pixel sizes: (a) original sonar images of different pixel sizes; (b) detection results.

Figure 20.

Figure 20

Detection accuracy of different pixel sizes.

Due to the influence of the seafloor environment, such as underwater acoustic channel, hydrological medium, and electromagnetic wave transmission, the imaging characteristics of sensor, high noise, and weak boundaries are commonly detected in the sensor images of detection region. The speckle noise has the most considerable influence on the detection of sonar image. In order to verify the detection accuracy of DS-CNNs in the noise background, the Gaussian noise, Poisson noise, multiplicative noise, and salt noise with a signal-to-noise ratio (SNR) of 35 are added to the original sonar image. The detection experiment of the sonar image with noise was performed using DS-CNNs, and the results are shown in Table 4 and Figure 21. It can be seen that the DS-CNNs still have a favorable detection effect on the sonar images under noise background, in which Poisson noise has the least influence on the detection effect and the average detection accuracy reaches 95.12%. Under the background of Gaussian noise, the detection effect is poor, but the average detection accuracy also reaches 93.28%. The detection results under the noise background show that DS-CNNs are robust to noises.

Table 4.

Detection accuracy of sonar image with different noise types.

Noise type SP (%) SE (%) Acc (%) Jacc (%) Dice (%)
Gaussian noise 78.62 81.26 93.28 83.35 84.57
Poisson noise 81.23 83.68 95.12 86.01 87.32
Multiplicative noise 79.16 82.31 94.36 84.53 85.13
Salt noise 78.91 82.06 93.81 83.98 84.37

Figure 21.

Figure 21

Detection results of different noise backgrounds: (a) original image; (b) Gaussian noise; (c) Poisson noise; (d) multiplicative noise; (e) salt noise.

Objective detection task is performed in randomly selected 1850 sonar images using DS-CNNs, and the 1850 sonar images include both objective and nonobjective types. The receiver operating characteristic (ROC) and area under curve (AUC) were used to evaluate the effect of the model on 1850 sonar images. The ROC and AUC of DS-CNNs are shown in Figure 22. It is analyzed from Figure 22 that the models have excellent stability in the sonar objective detection task. The detection accuracy of the sonar image with objective and nonobjective is more than 94.5%, indicating that the DS-CNNs are applicable to solve the problem of sonar objective detection.

Figure 22.

Figure 22

ROC and AUC of DS-CNNs.

5. Conclusion

Accurate detection of underwater sonar objective is an important task for underwater detection. In order to solve the shortcomings of the existing detection methods, this article proposed a novel network, DS-CNNs, for detection of underwater sonar objective. The innovations of DS-CNNs include the dilated separable convolution kernel, multisample pooling, and QPSO optimization algorithm. In the experiment process, it is found that the dilated separable convolution kernel can effectively extract the multiscale feature of the sonar objective; the multisampling pooling operation can reduce the loss of objective feature information; the QPSO optimization algorithm can speed up the model training speed and alleviate the overfitting phenomenon. To evaluate the effectiveness of the proposed method, DS-CNNs are used to perform objective detection experiments on sonar image datasets and compared with existing methods.

Experimental results show that the detection accuracy of DS-CNNs reaches 96.98%, which is better than other compared methods. Moreover, DS-CNNs have short model training time and few calculation parameters, so it has strong practicability. In the process of robustness evaluation, the proposed method still has high-objective detection accuracy for sonar images containing noise, indicating that DS-CNNs have strong robustness to noise and underwater environmental changes. In future work, we will concentrate on how to reduce its number of iterations and shorten its running time. In addition, we will investigate how to compress network model parameters and apply them to the practical application of underwater portable devices.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China under Grant 61671465.

Data Availability

The source code and sonar dataset used to support the findings of this study have been in the https://figshare.com/s/99620ba7a7bfff1400c2 repository.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

References

  • 1.Mizrahi M., Amy D., Rebecca W., et al. A systematic review of the socioeconomic factors that influence how marine protected areas impact on ecosystems and livelihoods. Society & Natural Resources. 2018;7(2):1–17. [Google Scholar]
  • 2.Postolache O., Girao P., Pereira M. Underwater acoustic source localization based on passive sonar and intelligent processing. Proceedings of the 2019 Instrumentation and Measurement Technology Conference; November 2019; Warsaw, Poland. pp. 2502–2521. [Google Scholar]
  • 3.Mariani P., Quincoces I., Haugholt K., et al. Range-gated imaging system for underwater monitoring in ocean environment. Sustainability. 2018;11(1):126–139. doi: 10.3390/su11010162. [DOI] [Google Scholar]
  • 4.Chen Z., Wang R., Ji W., et al. A novel monocular calibration method for underwater vision measurement. Multimedia Tools and Applications. 2019;78(9):19437–19455. doi: 10.1007/s11042-018-7105-z. [DOI] [Google Scholar]
  • 5.Yang S., Wang Y., Tian Y. Laser underwater image target detection based on Gabor filter. The International Society for Optical Engineering. 2009;12:18–36. [Google Scholar]
  • 6.Lianantonakis M., Petillot Y. R. Sidescan sonar segmentation using texture descriptors and active contours. IEEE Journal of Oceanic Engineering. 2007;32(3):744–752. doi: 10.1109/joe.2007.893683. [DOI] [Google Scholar]
  • 7.Elias F., George P., Maria G., et al. An automatic target detection algorithm for swath sonar backscatter imagery, using image texture and independent component analysis. Remote Sensing. 2016;8(5):373–381. [Google Scholar]
  • 8.Ganesh E. N., Kumaratharan N. Image segmentation based on Markov random field probabilistic approach. Proceedings of the 2019 International Conference on Communication and Signal Processing (ICCSP); April 2019; Chennai, India. [Google Scholar]
  • 9.Mignotte M., Collet C., Pérez P., Bouthemy P. Unsupervised segmentation applied on sonar images. Energy Minimization Methods in Computer Vision and Pattern Recognition. 2005;12(2):491–506. doi: 10.1007/3-540-62909-2_99. [DOI] [Google Scholar]
  • 10.Cacciola M., Calcagno S., Megali G., Morabito F. C., Pellicano D., Versaci M. FEA design and misfit minimization for in-depth flaw characterization in metallic plates with eddy current nondestructive testing. IEEE Transactions on Magnetics. 2009;45(3):1506–1509. doi: 10.1109/tmag.2009.2012691. [DOI] [Google Scholar]
  • 11.Collet C., Mignotte M., Perez P., et al. Sonar image segmentation using an unsupervised hierarchical MRF model. IEEE Transactions on Image Processing. 2000;9(7):1216–1231. doi: 10.1109/83.847834. [DOI] [PubMed] [Google Scholar]
  • 12.Yan S., Bo H., Peng L., et al. Side scan sonar image segmentation and synthesis based on extreme learning machine. Applied Acoustics. 2019;16(1):186–193. [Google Scholar]
  • 13.Min S. B., Chen X. J., Xie H. T., et al. Accurate segmentation of synaptic cleft with contour growing concatenated with a convnet. Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP); August 2019; Taipei, Taiwan. [Google Scholar]
  • 14.Angiulli G., Calcagno S., Versaci M., et al. Second-order parabolic equation to model, analyze, and forecast thermal-stress distribution in aircraft plate attack wing-fuselage. Mathematics. 2020;8(6):311–332. doi: 10.3390/math8040489. [DOI] [Google Scholar]
  • 15.Peng W. J., Guo H. T. Underwater sonar image segmentation based on snake model. Applied Mechanics and Materials. 2013;448–453:3675–3678. [Google Scholar]
  • 16.Wang X., Zhao J., Zhu B., et al. A side scan sonar image target detection algorithm based on a neutrosophic set and diffusion maps. Remote Sensing. 2018;10(2):185–295. doi: 10.3390/rs10020295. [DOI] [Google Scholar]
  • 17.Zhao J., Wang X., Zhang H., Hu J., Jian X. Side scan sonar image segmentation based on neutrosophic set and quantum-behaved particle swarm optimization algorithm. Marine Geophysical Research. 2016;37(3):229–241. doi: 10.1007/s11001-016-9276-1. [DOI] [Google Scholar]
  • 18.Guo Y., Wei L., Xu X., et al. A sonar image segmentation algorithm based on quantum-inspired particle swarm optimization and fuzzy clustering. Neural Computing and Applications. 2018;16(3):1–8. [Google Scholar]
  • 19.Duan J., He Y., Du B., Ghandour R. M. R., Wu W., Zhang H. Intelligent localization of transformer internal degradations combining deep convolutional neural networks and image segmentation. IEEE Access. 2019;7:62705–62720. doi: 10.1109/access.2019.2916461. [DOI] [Google Scholar]
  • 20.Zhang N., Donahue J., Girshick R., Darrell T. Part-based R-CNNs for fine-grained category detection. Proceedings of the 2014 European Conference on Computer Vision—ECCV 2014; September 2014; Zurich, Switzerland. [DOI] [Google Scholar]
  • 21.He K. M., Zhang X. Y., Ren S. Q., et al. Deep residual learning for image recognition. Proceedings of the 2016 the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); June 2016; Las Vegas, NV, USA. [Google Scholar]
  • 22.Song Y., Zhu Y., Li G., et al. Side scan sonar segmentation using deep convolutional neural network. Oceans. 2017;17(2):1–4. [Google Scholar]
  • 23.Valdenegro M. Object recognition in forward-looking sonar images with convolutional neural networks. Oceans. 2016;98(2):209–219. [Google Scholar]
  • 24.Zhu P., Isaacs J., Fu B., et al. Deep learning feature extraction for target recognition and classification in underwater sonar images. Proceedings of the 2017 IEEE 56th Annual Conference on Decision and Control (CDC); January 2017; Melbourne, Australia. pp. 2724–2731. [Google Scholar]
  • 25.Dzieciuch I., Gebhardt D., Barngrover C., et al. Non-linear convolutional neural network for automatic detection of mine-like objects in sonar imagery. Proceedings of the 4th International Conference on Applications in Nonlinear Dynamics; March 2016; Westminster, CO, USA. [Google Scholar]
  • 26.Chen J. L., Summers J. E. Deep convolutional neural networks for semi-supervised learning from synthetic aperture sonar (SAS) images. The Journal of the Acoustical Society of America. 2017;141(5):p. 3963. doi: 10.1121/1.4989020. [DOI] [Google Scholar]
  • 27.Jiao S., Zhao C., Xin Y., et al. Research on convolutional neural network model for sonar image segmentation. MATEC Web of Conferences. 2018;220:18–24. doi: 10.1051/matecconf/201822010004. [DOI] [Google Scholar]
  • 28.Williams D. P., Dugelay S. Multi-view SAS image classification using deep learning. Proceedings of the OCEANS 2016 MTS/IEEE Monterey; September 2016; Monterey, CA, USA. pp. 1–9. [Google Scholar]
  • 29.Wang Z., Zhang S. Sonar image detection based on multi-scale multi-column convolution neural networks. IEEE Access. 2019;7(1):160755–160767. doi: 10.1109/access.2019.2951443. [DOI] [Google Scholar]
  • 30.Bin P., Tian J. H., Qi Z., et al. Cascade convolutional neural network based on transfer-learning for aircraft detection on high-resolution remote sensing images. Journal of Sensors. 2017;16(1):1–14. [Google Scholar]
  • 31.Darwish A., Hassanien A. E., Das S. A survey of swarm and evolutionary computing approaches for deep learning. Artificial Intelligence Review. 2020;5(53):1767–1812. doi: 10.1007/s10462-019-09719-2. [DOI] [Google Scholar]
  • 32.Alejandro M., Lara-Cabrera R., Fuentes F. EvoDeep: a new evolutionary approach for automatic deep neural networks parameterizations. Journal of Parallel and Distributed Computing. 2018;3(117):180–191. [Google Scholar]
  • 33.Watanabe S., Shinozaki T. Structure discovery of deep neural network based on evolutionary algorithms. Proceedings of the 2015 IEEE International Conference on acoustics, speech, and signal processing, (ICASSP); 2015; Brisbane, Australia. [Google Scholar]
  • 34.Real E., Moore S., Selle A., et al. Large-scale evolution of image classifiers. Proceedings of the 2017 International Conference on Machine Learning (ICML); 2017; Ho Chi Minh City, Vietnam. pp. 2902–2911. [Google Scholar]
  • 35.Chung H., Shin K. Genetic algorithm-optimized multi-channel convolutional neural network for stock market prediction. Neural Computing and Applications. 2020;8(32):7897–7914. doi: 10.1007/s00521-019-04236-3. [DOI] [Google Scholar]
  • 36.Połap D., Marcin W., Wei W. Multi-threaded learning control mechanism for neural networks. Future Generation Computer Systems. 2018;2(87):16–34. [Google Scholar]
  • 37.Moayedi H., Tien D., Dounis A., et al. Novel nature-inspired hybrids of neural computing for estimating soil shear strength. Applied Sciences. 2019;9(16):4643–4650. doi: 10.3390/app9214643. [DOI] [Google Scholar]
  • 38.Ronneberger O., Fischer P., Brox T. U-net: convolutional networks for biomedical image segmentation. Proceedings of the 2015 International Conference on Medical Image Computing and Computer-Assisted Intervention; November 2015; Munich, Germany. [DOI] [Google Scholar]
  • 39.Long J., Shelhamer E., Darrell T. Fully convolutional networks for semantic segmentation. IEEE Transactions on Pattern Analysis & Machine Intelligence. 2014;39(4):640–651. doi: 10.1109/TPAMI.2016.2572683. [DOI] [PubMed] [Google Scholar]
  • 40.Badrinarayanan V., Kendall A., Cipolla R. SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2017;39(12):2481–2495. doi: 10.1109/tpami.2016.2644615. [DOI] [PubMed] [Google Scholar]
  • 41.Szegedy C., Liu W., Jia Y., et al. Going deeper with convolutions. IEEE Transactions on Pattern Analysis & Machine Intelligence. 2014;12(1):16–30. [Google Scholar]
  • 42.Howard A. G., Zhu M., Chen B., et al. MobileNets: efficient convolutional neural networks for mobile vision applications. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition; April 2017; Honolulu, HI, USA. pp. 163–177. [Google Scholar]
  • 43.Chen L. C., Papandreou G., Schroff F., et al. Rethinking atrous convolution for semantic image segmentation. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition; April 2017; Honolulu, HI, USA. pp. 56–73. [Google Scholar]
  • 44.Huang G., Liu Z., Laurens V. D. M., et al. Densely connected convolutional networks. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition; August 2016; Las Vegas, NV, USA. pp. 16–31. [Google Scholar]
  • 45.He K., Zhang X., Ren S., Sun J. Identity mappings in deep residual networks. Proceedings of the 2016 European Conference on Computer Vision; September 2016; Amsterdam, Netherlands. [DOI] [Google Scholar]
  • 46.Zhao H., Shi J., Qi X., et al. Pyramid scene parsing network. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition; August 2016; Las Vegas, NV, USA. pp. 152–163. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The source code and sonar dataset used to support the findings of this study have been in the https://figshare.com/s/99620ba7a7bfff1400c2 repository.


Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES