Skip to main content
PLOS One logoLink to PLOS One
. 2021 Feb 10;16(2):e0246093. doi: 10.1371/journal.pone.0246093

Fast semantic segmentation method for machine vision inspection based on a fewer-parameters atrous convolution neural network

Jian Huang 1, Liu Guixiong 1,*, Binyuan He 1
Editor: Jeonghwan Gwak2
PMCID: PMC7875430  PMID: 33566844

Abstract

Owing to the recent development in deep learning, machine vision has been widely used in intelligent manufacturing equipment in multiple fields, including precision-manufacturing production lines and online product-quality inspection. This study aims at online Machine Vision Inspection, focusing on the method of online semantic segmentation under complex backgrounds. First, the fewer-parameters optimization of the atrous convolution architecture is studied. Atrous spatial pyramid pooling (ASPP) and residual network (ResNet) are selected as the basic architectures of ηseg and ηmain, respectively, which indicate that the improved proportion of the participating input image feature is beneficial for improving the accuracy of feature extraction during the change of the number and dimension of feature maps. Second, this study proposes five modified ResNet residual building blocks, with the main path having a 3 × 3 convolution layer, 2 × 2 skip path, and pooling layer with ls = 2, which can improve the use of image features. Finally, the simulation experiments show that our modified structure can significantly decrease segmentation time Tseg from 719 to 296 ms (decreased by 58.8%), with only a slight decrease in the intersection-over-union from 86.7% to 86.6%. The applicability of the proposed machine vision method was verified through the segmentation recognition of the China Yuan (CNY) for the 2019 version. Compared with the conventional method, the proposed model of semantic segmentation visual detection effectively reduces the detection time while ensuring the detection accuracy and has a significant effect of fewer-parameters optimization. This slows for the possibility of neural network detection on mobile terminals.

Introduction

Semantic segmentation is a basic task in computer vision and is aimed at dividing a visual input into different semantically interpretable categories (that is, assigning a semantic label to each pixel of the image [1]). Although unsupervised methods, such as clustering, can be used for segmentation, their results do not necessarily possess semantics. Fig 1 presents the flowchart of the general method of semantic segmentation, including input, area generation, feature extraction, classifier, post-processing, and output segmentation results. The flow-segmentation method is composed of multiple independent algorithm modules, which are designed by experts; for example, the second-order pooling (SOP) [2], discriminative re-ranking of diverse segmentations (DRS) [3], unified detection and segmentation (UDS) [4] and simultaneous detection segmentation (SDS) [5].

Fig 1. Flowchart of the general semantic segmentation method.

Fig 1

Semantic segmentation has the best fine-grained identification and can distinguish the entity levels/backgrounds (centroid) of the components through indirect calculation. The method also has a clear physical definition and is suitable for machine-vision identification with high spatial resolution and reliability. The general machine-vision system belongs to a complex photoelectric system with high requirements of accuracy, real-time applicability, and repeatability [69]. A convolutional neural network (CNN) for semantic segmentation was applied to assign a semantic label to each pixel [10]. The CNN semantic segmentation method is derived from a single-step end-to-end CNN semantic segmentation model divided into multiple modules for avoiding extensive processing. However, the connection mode of various modules directly affects CNN, while rendering of the model increases dramatically. The required vast storage and computational overhead severely limit the application of CNN in the field of low power consumption for visual detection.

Atrous convolution was first presented in the field of image segmentation. In this method, the image is input into the network to extract features through CNN. Subsequently, the image scale is reduced by pooling while the receptive field is increased. Although the CNN semantic segmentation method is a single-step end-to-end semantic segmentation, it is not divided into multiple modules. Nevertheless, the connection mode of various modules directly affects CNN.

Owing to the rapid development in hardware and computing power, the fewer-parameters network structure can effectively solve the problem of training and prediction efficiency due to the modeling complexity. The prediction efficiency is mainly based on the model storage and prediction speed. The storage problem is mainly due to the presence of hundreds of networks, which in turn indicates dealing with a great number of weighting parameters; the storage of a large number of weighting parameters demands for a large device memory. Moreover, the speed problem is mainly based on practical applications. The atrous convolution architecture eliminates part of the CNN pooling layer while replacing the convolutional layer with a cascade or parallel atrous convolutional layer, enabling the analysis of the feature map at multiple arbitrary scales, and thus significantly improving the segmentation accuracy [1113] and providing the possibility of detecting applications in the field of low power consumption. For obtaining a more accurate and faster fewer-parameters model as well as a method to achieve online machine vision and identification, in this study, the weight optimization technology was investigated based on the downsampling atrous convolution architecture-network model. This paper first presents a discussion of the associated indicators, including accuracy, computing speed, space occupation, and training speed, and then the modeling of the fewer-parameters of atrous convolutional architecture, by selecting ηseg and ηmain as the basic architecture. Next, an analysis of the fewer-parameters mechanism of downsampling residual blocks to improve the residual network (ResNet) by using dmain and dmain is presented to show that a ResNet structure can be modified by downsampling residual blocks.

The main contributions of this article are as follows:

  1. Based on the parameters of the dense convolutional architecture network, including dense prediction network structure ηseg, main network structure ηmain, and main network depth dmain, we designed a fewer-parameters optimization mathematical model according to the atrous convolution architecture network.

  2. We analyze ResNet downsampling residual construction blocks and point out that improving input utilization during downsampling is beneficial for improving the accuracy of the ResNet. The fewer-parameters optimization of ResNet and dmain could be accomplished based on the downsampling of residual blocks.

  3. We adopted the downsampled residual block to improve the ResNet structure after comparing the modified structure with the 3 × 3 main path and ls = 2 convolutional layers; 2 × 2 skip path and ls = 2 pooling layers; 3 × 3 main path, 2 × 2 skip path, and ls = 2 pooling layers. The application of machine-vision segmentation in the 2019 version of the China Yuan (CNY) showed that the proposed modified network benefits from the increase of the mean intersection over union (IoU¯) and the decrease in the segmentation time, Tseg.

The remainder of this article is organized as follows. Section 2 presents the recent studies related to our study, including a multi-empty convolutional network structure and RestNet representative network. The proposed method is presented in Section 3, with the focus on three aspects: 1) a discussion related to evaluation indicators; 2) the fewer-parameters optimization modeling and selection of the basic architecture of ηseg and ηmain; and 3) the improved fewer-parameters mechanism of ResNet and dmain based on the downsampling of residual blocks as well as the modification of the ResNet structure based on this downsampling. The experimental application of the proposed method is presented in Section 4 and the conclusions are drawn in Section 5.

Related work

The current CNN end-to-end semantic segmentation method can process images of any resolution, identify multiple objects at the pixel scale, and output multivalue maps without resolution loss [14]. This process can be integrated into machine vision, thus forming an artificial intelligence method with great generalization in precision measurement and analysis. Unlike the codec architecture that retains the pooling layer as the encoder, the atrous convolution architecture establishes other techniques to achieve semantic segmentation. Atrous convolution removes part of the pooling layer and replaces it with the convolutional layer and the fully connected layers with the atrous convolution to maintain a high resolution of the feature map. ResNet is the ILSVRC 2015 champion network [15]. The main idea of the residual block structure is to increase the direct connection channel in the network; that is, the idea of the highway network, which effectively solves the problem of numerous network layers, causing gradient dispersion or gradient explosion problems. We analyzed the research progress on the atrous convolution and ResNet network structure to provide the feasibility and theoretical basis for subsequent research.

CNN-based semantic segmentation

Typical CNN-based semantic segmentation networks include fully convolutional networks (FCN) [16], SegNet [17], and pyramid scene parsing network (PSPNet) [18].

FCN [16] is an end-to-end semantic segmentation network proposed by Jonathan Long et al. in 2014. Its distinguishing characteristic is the conversion of a fully connected layer into a convolutional layer. It is capable of processing images of any resolution, successfully overcoming the limitation of a fully connected layer—that is, only being capable of processing images of a specific resolution. However, FCN suffers from certain problems, including the loss of details, smoothening of complex boundaries, and poor detection sensitivity in the case of small object. FCN gain a mIoU accuracy 62.2% on PASCAL VOC 2012. SegNet [17] is an efficient, real-time, and end-to-end semantic segmentation network proposed by Alex Kendall et al. in 2015. It identifies a one-to-one correspondence between the decoder and encoder by using the maximum pooling index of the encoder to perform non-linear upsampling to form a sparse feature image. Following this, a dense feature map is generated via convolution. However, the accuracy of SegNet is lower than that of FCN, and it also suffers from the issue of boundary misdetection. SegNet gain a mIoU accuracy 59.9% on PASCAL VOC 2012. PSPNet [18] was proposed by Zhao Hengshuang in 2017, PSPNet exploiting the capability of global context information by different-region-based context aggregation through a novel pyramid pooling module together with the proposed pyramid scene parsing network. PSPNet’s global prior representation is effective to produce good quality results on the scene parsing task, while PSPNet provides a superior framework for pixel-level prediction tasks. A single PSPNet yields mIoU accuracy 85.4% on PASCAL VOC 2012.

Atrous CNN-based semantic segmentation

Typical atrous CNN-based semantic segmentation networks include DeepLab [19] and dilated convolution network (DCN) [20], DeepLab v2 [21] and DeepLab v3 [22].

DeepLab [19] was proposed by the University of California and Google in 2015. Fig 2 shows the structure of the DeepLab model, in which the last two pooling layers of CNN are removed and α = 2 and 4 atrous convolution layers are used to replace a convolutional layer and a fully connected layer, respectively. Thus, the network outputs a 1/8 original-resolution feature map (which is more than that of the feature map outputted by FCN) and then upsamples it to restore the original resolution by using a conditional random field to improve boundary prediction accuracy [23]. DeepLab proposes a fundamental paradigm of cascaded atrous convolution architecture. The segmentation of a 480 × 360 image by using DeepLab requires 110.06 ms, which is 73.95% and 65.29% faster than the SegNet and FCN methods of the codec architecture. Moreover, its IoU¯ was tested at 67.1% for VOC2012.

Fig 2. DeepLab model structure diagram.

Fig 2

In 2016, Princeton University and Intel Corporation proposed DCN [20]. The atrous algorithm includes signal decomposition at multiple scales, and atrous convolution is applied at multiple scales to extract feature maps. Fig 3 depicts the structure of the DCN model, comprising a set of cascaded context networks with seven 3 × 3 atrous convolutions and one 1 × 1 convolution based on DeepLab; the model uses multiscale information to improve accuracy. The DCN has an IoU¯ of 67.6% in VOC2012, 0.5% higher than DeepLab.

Fig 3. DCN model structure diagram.

Fig 3

In 2016, the University of London and Google proposed DeepLab v2 [21]. Fig 4 shows the model of the DeepLab v2 structure using the atrous spatial pyramid pooling (ASPP) instead of the fully connected layer in CNN. ASPP is composed of four atrous convolutions with atrous ratios of 6, 12, 18, and 24, used in VGG and ResNet obtained an IoU¯ of 71.6% and 79.7%, respectively, in VOC 2012. DeepLab v2 can consistently detect large object areas.

Fig 4. Model of DeepLab v2 structure.

Fig 4

In 2017, Google proposed DeepLab v3 [22], as illustrated in Fig 5. This model improved the ASPP module by introducing a 1 × 1 convolutional layer and global pooling layer for maintaining features. It modifies the feature fusion method that stiches the output feature map in the third dimension, and then upsamples it to restore the resolution. DeepLab v3 could obtain an IoU¯ of 86.9% in the VOC2012 dataset.

Fig 5. Model of DeepLab v3 structure.

Fig 5

ResNet network

He et al. [15] proposed ResNet, in which the “degeneration” problem is solved through the residual block model. The main idea of the residual block structure in ResNet is to add the direct connection channel into the network, i.e., the idea of a highway network. Fig 6 shows the structure of the ResNet network. After the image is input, the short connection of the dimension matching converts from a dotted line into a solid line. When the dimensions do not match, two equivalent mapping methods can be selected, i.e., directly adding zero to increase the dimension or multiplying the W matrix to project into a new space. This allows the network to theoretically always remain in an optimal state, and thus the performance will not deteriorate with depth.

Fig 6. Structural model of ResNet.

Fig 6

When the model becomes more complicated, some problems could arise; for instance, the accuracy could drop rapidly after saturation, resulting in higher training errors or the stochastic gradient descent (SGD) optimization could become more difficult. To solve these problems, the residual structure is used such that the depth of the network model is unlimited in a large range (currently more than 1000 layers). The error rate of ResNet on Top-5 is 4.49%, and the number of network model parameters is less than those used in VGGNet with outstanding results. ResNet is representative of the current deepening model but its deepening increases the network size, and thus more storage space is required.

Our preliminary research [24] found that, replacing the FCN’s VGG16 backbone with ResNet50, IoU¯ on VOC2012 increase from 62.6% to 83.6%.Yu et al. [25] replacing the DeepLab’s VGG16 backbone with ResNet50, IoU¯ on VOC2012 increase from 67.1% to 75.6%. Lu et al. [26] replacing the DCN ‘s VGG backbone with ResNet50, IoU¯ on VOC2012 increase from 67.6% to 81.0%.

The atrous convolutional network has a greater advantage in improving the accuracy of boundary prediction, and it has improved the performance of semantic segmentation in recent years. Currently, ResNet network has the advantages of high accuracy in classification and an excellent IoU¯ combined with multiple semantic segmentation networks. Therefore, further research on the fewer-parameters optimization of the model of the atrous convolution network architecture can fulfill the requirements of accuracy and its speed in the online machine vision detection and identification.

Proposed method

The proposed method considers the pixel accuracy PA and mIoU IoU¯ [27]. Segmentation time Tseg is defined as the time the algorithm requires to segment the image, and IoU¯ is defined as the degree of overlap between the segmentation result and its true value. Under the condition that IoU¯ satisfies the minimum IoUmin meeting, the calculation speed index, Tseg, must be optimized. A fewer-parameters optimization model is based on the accuracy requirements of IoUmin, IoU¯, and Tseg. The method combines the dense prediction network structure ηseg, backbone network structure ηmain, and depth dmain. The basic architecture with relatively better ηseg and ηmain was selected. Improvements in the downsampling of the residual building blocks of the main path and skipped path are proposed to optimize ηmain and dmain, and thus obtain a fewer-parameters and optimized atrous convolution architecture network.

Let k be the number of classes of the object to be detected with machine vision. The semantic segmentation model requires the identification of k + 1 labels, where L = {l0,l1lk}, including the background. Let the number of li type of the pixels recognized as the lj class be pij (thus, the number of li type of the pixels recognized as the li class is pii). Then, PA and mean IoU¯ can be calculated as follows:

PA=i=0kpiii=0kj=0kpij, (1)
IoU¯=i=0kpiij=0kpij+j=0kpjipii. (2)

Fewer-parameters optimization modeling of atrous convolution architecture and selection of basic architecture with ηseg and ηmain ResNet networks

The accuracy of pixel classification and the amount of segmentation can be measured by using IoU¯ and Tseg, respectively. The larger the value of IoU¯, the higher is the accuracy of pixel classification. The larger the value of Tseg, the larger is the amount of segmentation. In the fewer-parameters optimization of the semantic segmentation network, the aim of fast semantic segmentation technology is to reduce Tseg under the condition that IoU¯ meets the requirements of IoUmin.

We can build an atrous convolutional neural network by determining backbone network structure ηmain, backbone network depth dmain, and dense prediction network structure ηseg. Therefore, the IoU¯ and Tseg of this network can be expressed as IoU¯(ηseg,ηmain,dmain) and Tseg(ηseg,ηmain,dmain), respectively. The mathematical model of the fewer-parameters optimization based on atrous convolution architecture network can be derived as follows:

{minTseg(ηseg,ηmain,dmain)s.t.IoU¯(ηseg,ηmain,dmain)IoUmin, (3)

where ηseg, ηmain, and dmain are the optimization parameters.

To optimize ηseg, ηmain and dmain in Eq (3), they are first combined. In addition, a basic architecture with relatively better ηseg and ηmain is selected by comparing the values of IoU¯ and Tseg of all networks. In this study, the authors used commonly used semantic segmentation networks, such as FCN [16], improved FCN [24], PSPNet [18], and DeepLabv3 [22], available from the representative computer vision deep learning software package, Amazon GluonCV [28]. The IoU¯ was obtained from the typical semantic segmentation dataset, PASCAL VOC [1]. The images were segmented using the GeForce GTX 1080Ti GPU hardware environment at Tseg to obtain segmentation results with the same resolution of 1280 × 1024. Table 1 lists the performance indexes of the main semantic segmentation network [28], where the relatively better indexes are represented in squares.

Table 1. Parameters and indicators of the main semantic segmentation CNN.

Number CNNs Dense prediction network ηseg Backbone ηmain Depth dmain PASCAL VOC IoU¯ execution time Tseg (ms)
1 FCN FCN VGG 16 62.2% 347
2 Modified FCN FCN ResNet 101 83.6% 404
3 PSPNet PPM ResNet 101 85.1% 510
4 DeepLabv3 ASPP ResNet 101 86.2% 495
5 DeepLabv3 ASPP ResNet 152 86.7% 646

The following points should be noted:

  1. Among networks 1, 2, 3, and 4 in Table 1, with the same ηseg, the IoU¯ values for ηmain = ResNet is better than the one for ηmain = VGG.

  2. ηmain and dmain are the same in networks 2, 3, and 4. The IoU¯ for ηseg = ASSP is better than those for ηseg = PPM or ηseg = PSPNet.

  3. ηseg and ηmain are the same in networks 4 and 5. The Networks 5 IoU¯ shows a slight improvement, and Tseg increases significantly when dmain increases by approximately 50%. Therefore, in this study, the basic architectures of ηseg = ASSP and ηmain = ResNet were selected.

Improvement of ResNet based on downsampling residual building blocks and fewer-parameters mechanism of dmian

Fig 7 shows the parameter and index of the atrous convolution architecture network for semantic segmentation. The figure also shows the ηmain hidden-layer type, building block structure, and layer operation parameter of ResNet. dmain represents the number of hidden layers. Fig 8 shows the ResNet model and its bottleneck building block structure. The ResNet backbone network is composed of a 7 × 7 convolutional layer, 3 × 3 maximum pooling layer, multiple residual building blocks, and downsampled residual building block. Each stage (i.e., the convolutional layer, pooling layer, and downsampling residual building block) reduces the resolution of the feature map by 1/2. The main and skipped paths of the residual and downsampled residual building blocks are connected in series with at most three hidden layers (depth = 3).

Fig 7. Parameter and index of atrous convolution architecture network for semantic segmentation.

Fig 7

Fig 8. ResNet model and bottleneck building block structure.

Fig 8

a) ResNet model structure, b) bottleneck block structure, and c) downsampled building block structure.

The residual unit in the neural network is composed of multiple hidden layers and skip connections. As shown in Fig 9, the input is an n-dimensional column vector X; the i-th hidden layer has a nci×nci1 matrix, nc0 with weights Wi. The output of the nci-dimensional column vector Hi. Moreover, the residual unit has three hidden layers (i =3), outputs of the nc3-dimensional column vector, Y, are represented as follows [15]:

Y=fNet[X,(W1,W2,W3)]+X. (4)

Fig 9. Residual learning function of ResNet bottleneck building block.

Fig 9

Suppose the actual value is YGT. The fitting target of the three hidden layers in the residual unit is the residual value between YGT and X:

fNet[X,(W1,W2,W3)]=YGTX. (5)

However, (4) and (5) are satisfied only when the YGT and X dimensions are consistent (nc3=nc0). When nc3nc0, the X dimension must be transformed to be consistent with the YGT dimension. The hidden layer with the Wskip as the nc3×nc0 matrix can be added to the skip path to make WskipX as the nc3-dimensional column vector. The residual unit operation is as follows:

Y=fNet[X,(W1,W2,W3)]+WskipXY=fNet[X,(W1,W2,W3)]+fskip(X,Wskip). (6)

ResNet downsampling residual building blocks [15] satisfy (6).

The output of the ResNet downsampling residual building block, fNet[X,(W1,W2,W3)], differs from the X dimension. Output Y and input X represent feature maps, the numbers and sizes of which vary. In the process of changing the number and size of feature maps, if the coefficient of the pixels of the feature map constantly equals 0, the image features are not used; this reduces the accuracy of feature extraction.

Note that 75% coefficients in W1, Wskip of fNet(X), fskip(X) are 0. The image-feature utilization is only 25% for input X.

Once the downsampling residual building block is optimized, the proportion of input-image features involved in the calculation is improved as much as possible during the change of the number and size of feature maps; this helps to improve the accuracy of feature extraction. Thus, the requirement of IoUmin is satisfied, and dmain and Tseg are reduced, achieving the improvement of ResNet based on downsampling residual building blocks and fewer-parameters optimization of dmain. The fewer-parameters optimization of the atrous convolution architecture network was then performed (which is the contribution of this article).

Improved ResNet structure based on downsampled residual building blocks

Based on the hidden-layer unit function and image-feature utilization listed in Table 2, the reasonable use of these hidden-layer units can help construct optimized ResNet downsampled residual building blocks.

Table 2. Function and image-feature utilization of general hidden layer structure.

Hidden layer structure Stride ls Convolutional Downsample Change channel number Image-feature utilization
1×1 Convolutional layer ls = 1 —— —— 100%
ls = 2 —— 25%
3×3 Convolutional layer ls = 1 —— 100%
ls = 2 100%
Pooling layer ls = 2 —— —— 100%

1) Improved structure of the main path with the 3 × 3 convolutional layer when ls = 2.

Fig 10 shows the modified structure of the downsampled building block with stride 2, 3 × 3 convolutional layers. In the improved scheme, the skip path does not change while the main path changes. The ls = 2 downsampling is moved from the 1 × 1 convolutional layer to the 3 × 3 convolutional layers. Table 2 shows that the dimension of output fNet(X) of the main path does not change in the improved structure. Moreover, the image-feature-utilization rate of the main path increases from the original 25% × 100% × 100% to 100% × 100% × 100%.

Fig 10.

Fig 10

Modified structure of the main path with the downsampled building block and stride 2 of 3×3 convolutional layers: a) Original structure and b) modified structure.

2) Improved structure of the skip path with 2 × 2 pooling layer and ls = 2

Fig 11 shows the downsampled building block with the modofoed structure of the skip path of 2 × 2 pooling layers. In the improved scheme, the main path does not change, while the skip path changes with the addition of a 2 × 2 pooling layer. The ls = 2 downsampling moves from 1 × 1 convolution layer to the added 2 × 2 pooling layer. Table 2 shows that the dimension of the skip-path output, fskip(X), does not change in the improved structure. The rate of the image-feature utilization of the skip path increased from the original 25% to 100% × 100%.

Fig 11.

Fig 11

The downsampled building block with the modified structure of the 2 × 2-pooling-layer skip path: a) modified structure I and b) modified structure II.

3) The improved structures of the main path with the 3 × 3 convolution layer and the skip path with 2 × 2 pooling layer ls = 2.

Fig 12 shows the downsampled building blocks of the main path with stride-2 of 3 × 3 convolutional layers and the modified structure of the skipped path with 2 × 2 pooling layers. The improvement scheme combines the improved structures of Figs 10 and 11. Table 2 shows that the dimensions of the outputs of both the main path [fNet(X)] and skip path [fskip(X)] do not change in the improved structure. The rate of the image-feature utilization of the main path increases from the original 25% × 100% × 100% to 100% × 100% × 100%. The rate of the image-feature utilization of the skip path increases from the original 25% to 100% × 100%. Figs 10(B), 11(A), 11(B), 12(A) and 12(B) show the improved ResNet structure based on downsampled residual building blocks proposed in this paper.

Fig 12.

Fig 12

Downsampled building block of the main path with stride-2 3 × 3 convolutional layers and modified skip path with 2 × 2 pooling layers: a) modified structure I and b) modified structure II.

Experiments and applications

Improved model on PASCAL VOC segmentation task

We replace the backbone of DeepLabv3 with the five types of improved structure to obtain the DeepLabv3 improved model. Evaluate the DeepLabv3 improved model with its IoU¯ of the PASCAL VOC dataset [1] segmentation task. In the PASCAL VOC segmentation task, IoUmin = 85%. We verified the accuracy of the IoU¯ index. Segmentation time Tseg was obtained for a 1280 × 1024 image, and the comparison of the IoU¯ and Tseg of the network model of the atrous convolution architecture before and after improving is presented in Table 3.

Table 3. IoU and segmentation time of DeepLab v3 with a modified ResNet in PASCAL VOC segmentation task.

Method Backbone ηmain Depth dmain = 50 Depth dmain = 101 Depth dmain = 152
IoU¯ Tseg IoU¯ Tseg IoU¯ Tseg
1 DeepLab v3 [22] ResNet [15] 84.7% 285 ms 86.2% 495 ms 86.7% 719 ms
2 DeepLab v3* ResNet v2 [29] 84.4% 276 ms 85.9% 469 ms 86.7% 677 ms
3 DeepLab v3* stride 2 3×3 conv layers mainpath modified structure 85.0% 288 ms 86.7% 474 ms 87.2% 680 ms
4 DeepLab v3* 2×2 pooling layers skippath modified structure I 85.4% 292 ms 87.1% 492 ms 87.6% 691 ms
5 DeepLab v3* 2×2 pooling layers skippath modified structure II 85.4% 299 ms 87.0% 502 ms 87.6% 704 ms
6 DeepLab v3* stride 2 3×3 conv layers mainpath and 2×2 pooling layers skippath modified structure I 86.6% 296 ms 88.1% 513 ms 88.2% 698 ms
7 DeepLab v3* stride 2 3×3 conv layers mainpath and 2×2 pooling layers skippath modified structure II 86.5% 303 ms 88.1% 523 ms 88.2% 711 ms

From Table 3, we can draw the following conclusions:

1) Replacing the backbone of DeepLabv3 with the five types of improved structure proposed in this paper, the accuracy of DeepLabv3 is improved under the same dmain situation. For example, depth dmain = 50, and each improved structure shows increase in the IoU¯ by 85.0%, 85.4%, 85.4%, 86.6%, and 86.5%. However, the values of Tseg also show sight increases from 285 to 288, 292, 299, 296, and 303 ms. This shows that improving the downsampled residual building block structure based on the improvement in the utilization of image features is effective for increasing IoU¯, and the adoption of the modified structure shown in Fig 12(A) is relatively better for improving the model structure.

2) The improvement in the downsampled residual building block structure produces an increase in IoU¯, while reducing dmain and Tseg. For the same IoUmin, after the improved downsampled residual building block structure is used, Tseg is reduced, i.e., to achieve fewer-parameters optimization of the network model relying on the atrous convolution architecture. For example, in Table 3, when ResNet + ASPP is used, with dmain = 101, we obtained IoU¯ = 86.2%, Tseg = 495 ms. In additio, when the structure in Fig 12(A) is used to improve the structure with ASPP, for dmain = 50, we obtained IoU¯ = 86. 6% and Tseg = 296 ms. Thus, we observed improvements in IoU¯ and Tseg indicators, and the fewer-parameters optimization effect depending on the network model of the atrous convolution architecture is evident. For example, when ResNet + ASPP is used, for dmain = 152, we obtained IoU¯ = 86.7% and Tseg = 719 ms. Morevover, when using the improved structure with ASPP (Fig 12(A)), for dmain = 50, IoU¯ = 86.7% but Tseg was reduced to 58.8%.

Fig 13 shows semantic segmentation result of the modified DeepLabv3 in the PASCAL VOC task.

Fig 13. PASCAL VOC semantic segmentation by modified DeepLabv3.

Fig 13

Improved model on CITYSCAPES pixel-level semantic labeling task

We trained each DeepLabv3 (with ResNet50 or modified ResNet50) on the pixel-level semantic labeling task of the CITYSCAPES dataset [30]. To assess performance, CITYSCAPES rely on the PASCAL VOC intersection-over-union metric IoU. Owing to the two semantic granularities, i.e. classes and categories, CITYSCAPES report two separate mean performance scores: IoUclass and IoUcategory. In either case, pixels labeled as void do not contribute to the score. We also report the segmentation time of each network run on a GeForce GTX 1080Ti GPU and an Intel i7-5960X CPU. Table 4 presents the performances of different methods on the CITYSCAPES pixel-level semantic labeling task. Table 5 presents the individual classes IoU of different methods in the CITYSCAPES pixel-level semantic labeling task. Fig 14 shows some CITYSCAPES pixel-level semantic labeling results obtained with the DeepLabv3 with different backbone.

Table 4. Performances of different methods on the CITYSCAPES pixel-level semantic labeling task.

Method IoUclass(%) IoUcategory(%) Segmentation time Tseg(ms)
DeepLab [19] 64.5 83.7 ——
DeepLab v2 [21] 70.4 86.4 ——
DeepLab v3 with ResNet50 [22] 78.7 89.8 494ms
DeepLab v3 with modified ResNet50 [ours] 79.9 91.6 497ms

Table 5. Individual classes IoU of DeepLabv3 with different backbone in CITYSCAPES pixel-level semantic labeling task.

Method road sidewalk building wall fence pole traffic light traffic sign vegetation terrain
DeepLab v3 with ResNet50 [22] 98.6% 86.4% 92.8% 52.4% 59.7% 59.6% 72.5% 78.3% 93.3% 72.8%
DeepLab v3 with modified ResNet50 [ours] 98.7% 87.0% 93.5% 57.9% 60.4% 70.9% 77.9% 81.4% 93.7% 72.8%
Method sky person rider car truck bus train motor cycle bicycle average
DeepLab v3 with ResNet50 [22] 95.5% 85.4% 70.0% 95.7% 75.4% 84.1% 75.1% 68.7% 75.0% 78.7%
DeepLab v3 with modified ResNet50 [ours] 95.6% 87.9% 75.3% 96.1% 65.8% 80.5% 78.7% 72.8% 70.8% 79.9%

Fig 14. CITYSCAPES pixel-level semantic labeling by modified DeepLabv3.

Fig 14

From Tables 4 and 5, we can draw the following conclusions:

1) The modified ResNet effectively improved DeepLabv3 performance in CITYSCAPES pixel-level semantic labeling task. Replacing the DeepLabv3’s ResNet50 backbone with our modified ResNet50 (stride 2 3×3 conv layers mainpath and 2×2 pooling layers skippath modified structure I), the IoUclass is improved from 78.7% to 79.9%, and the IoUcategory is improved from 89.8% to 91.6%. The segmentation time Tseg increased slightly from 494ms to 497 ms.

2) The modified ResNet also improved most of the individual classes IoU of the DeepLabv3. Especially, the wall, pole, traffic light, and rider IoU have increased by more than 5%. The traffic sign, person, train, and motor cycle IoU have increased by more than 2%.

Improved model on machine vision banknote anti-counterfeit segmentation

The following experiment was conducted using the 2019 version of the CNY machine vision segmentation and recognition application. In this task, the quantitative value for IoUmin was 90%. The experimental model includes an improved ResNet structure before and after the atrous convolution architecture network. In the DeepLab v3 semantic segmentation model, the dense prediction network was set as ASPP, the backbone network was replaced with various ResNet structures, the entire semantic segmentation network was trained with the 2019 version of the CNY image dataset. We verified the accuracy of ηseg and Tseg.

1) The machine vision system adopts an industrial camera, MV-CA013-10GC, with an MVL-HF2528M-6MP industrial lens, a bar light source with FOV = 18.33°, and an imaging resolution of 1280 × 1024.

2) For the image-dataset construction, the object plane is perpendicular to the optical axis. The object distance is 400 mm and various denominations, with 25 pieces of front and back renminbi, were collected with various angle. The 2019 version of the CNY image dataset (200 image) comprises the serial numbers on the CNY, magnetic optical variable ink (OVMI), security line denomination numbers, five patterns of visual features, etc..

3) Hardware condition: The hardware is the same as that used in the aforementioned experiment: GeForce GTX 1080Ti GPU.

Table 6 presents the IoU¯, execution time, and APIoUT=0.1 of DeepLab v3 with a modified ResNet in the segmentation task of CNY anti-counterfeit features. Fig 15 shows CNY anti-counterfeit feature segmentation with diffuse reflection machine vision and modified DeepLab v3.

Table 6. IOU and execution time of Deeplab v3 with a modified ResNet in CNY anti-counterfeit feature-segmentation task.

Method Backbone ηmain Depth dmain = 50 Depth dmain = 152
IoU¯ Tseg IoU¯ Tseg
1 DeepLab v3 [22] ResNet [15] 90.3% 287 ms 91.3% 722 ms
2 DeepLab v3* ResNet v2 [29] 89.9% 282 ms 91.1% 714 ms
3 DeepLab v3* stride 2 3×3 conv layers mainpath and 2×2 pooling layers skippath modified structure I 91.3% 295 ms 92.5% 735 ms

Fig 15. CNY anti-counterfeit features segmentation by modified DeepLabV3.

Fig 15

From Table 6, we can draw the following conclusions: The modified ResNet effectively improved DeepLabv3 performance in CNY anti-counterfeit feature-segmentation task. Replacing the DeepLabv3’s ResNet backbone with our modified ResNet (stride 2 3×3 conv layers mainpath and 2×2 pooling layers skippath modified structure I), the accuracy IoU¯ is improved under the same dmain situation. For example, in depth dmain = 50, our modified ResNet make IoU¯ increase from 90.3% to 91.3%, and Tseg show sight increases from 287 ms to 295 ms.

The conclusions drawn in Table 6 are consistent with the conclusions drawn for the PASCAL VOC dataset (Table 3). That is, the improved downsampled residual building block structure based on the improvement of the image-feature-utilization rate is helpful for improving IoU¯. Moreover, by adopting the improved structure with relatively better IoU¯ and Tseg, the fewer-parameters optimization of the network model of the atrous convolution architecture is evident.

Conclusions

In this study, fewer-parameters optimization was analyzed according to the atrous convolution network architecture model. For this, we selected ASPP and ResNet as the replacements for ηseg and ηmain basic architectures, respectively. Our analysis indicates that the optimization of the downsampled residual block increases the number of feature maps and the proportion of input image features used as much as possible. Specifically, in the application of the number and dimension of feature map during the change of the size of the input image, the accuracy of feature extraction is improved.

The proposed method innovatively modified the structures of the main path with a 3 × 3 convolution layer, 2 × 2 skip path, and ls = 2 pooling layer, with respect to five modified ResNet structures based on the downsampled residual building blocks, and thus improved the utilization of image features. As shown in the simulation experiments, under the same ηseg = ASPP, by using the modified ResNet instead of the conventional ResNet, the dmain changed from 152 to 50, enabling IoU to change from 86.7% to 86.6% and Tseg to change from 719 to 296 ms (decreased 58.8%). The fewer-parameters optimization of the atrous convolution architecture network model was also verified in the application of machine-vision segmentation recognition of 2019 version of CNY with an evident fewer-parameters optimization effect.

We have proposed the modified structure of the downsampled residual blocks to improve the utilization of image features that showed evident fewer-parameters optimization effects. The next step is to study the modified structure of more complicated downsampled residual blocks in networks such as ResNext and ResNest.

Data Availability

The PASCAL VOC Dataset used to support the findings of this study is available at http://host.robots.ox.ac.uk/pascal/VOC/. The CITYSCAPES Dataset used to support the findings of this study is available at http://cityscapes-dataset.com. The DeepLab v3 pretrained model are the implement of GLUON-CV, and available at https://gluon-cv.mxnet.io/model_zoo/segmentation.html. We provide the DeepLab v3 with modified ResNet50, and a program (*.ipynb) to use it on different image of PASCAL VOC, CITYSCAPES or CNY. https://github.com/HJ0116/utilization_semantic_segmentation

Funding Statement

This work was supported in part by the Key-Area Research and Development Program of Guangdong Province, China under Grant 2019B010154003, and in part by the Guangzhou Science and Technology Plan Project under Grant 201802030006.

References

  • 1.Everingham M, Eslami SMA, Van Gool L, Williams CKI, Winn J, Zisserman A. The PASCAL Visual Object Classes Challenge: A Retrospective. Int J Comput Vis. 2015;111(1):98–136. 10.1007/s11263-014-0733-5 WOS:000348345400006. [DOI] [Google Scholar]
  • 2.Carreira J, Caseiro R, Batista J, Sminchisescu C, editors. Semantic Segmentation with Second-Order Pooling. Computer Vision—ECCV 2012, Pt VII; 2012; Berlin: Springer-Verlag Berlin. [Google Scholar]
  • 3.Yadollahpour P, Batra D, Shakhnarovich G, Ieee. Discriminative Re-ranking of Diverse Segmentations. 2013 Ieee Conference on Computer Vision and Pattern Recognition; New York: Ieee; 2013. p. 1923–30.
  • 4.Dong J, Chen Q, Yan SC, Yuille A. Towards Unified Object Detection and Semantic Segmentation. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T, editors. Computer Vision—Eccv 2014, Pt V. Lecture Notes in Computer Science. 8693 Cham: Springer International Publishing Ag; 2014. p. 299–314. [Google Scholar]
  • 5.Hariharan B, Arbelaez P, Girshick R, Malik J. Simultaneous Detection and Segmentation. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T, editors. Computer Vision—Eccv 2014, Pt Vii. Lecture Notes in Computer Science. 8695. Cham: Springer International Publishing Ag; 2014. p. 297–312. [Google Scholar]
  • 6.Perez L, Rodriguez I, Rodriguez N, Usamentiaga R, Garcia DF. Robot Guidance Using Machine Vision Techniques in Industrial Environments: A Comparative Review. Sensors. 2016;16(3):26 10.3390/s16030335 WOS:000373713600152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Huang J, Liu GX. The Development of CNN-based Semantic Segmentation Method for Machine Vision Detection [J]. Laser Journal., 2019,40(05):10–16. (in Chinese). [Google Scholar]
  • 8.Liu GX, He BY, Wu JF, et al. Intelligent Detection Technology for Local Characteristics of Chassis Standard Component Assembly Quality [J]. CHINA MEASUREMENT&TEST. 2019,45(07):1–10. (in Chinese). [Google Scholar]
  • 9.Liu GX, Liu SYA, Wu JF, et al. Machine vision object detection algorithm based on deep learning and application in banknote detection. CHINA MEASUREMENT&TEST. 2019,45(05):1–9. (in Chinese). [Google Scholar]
  • 10.Yu JM, Kim D, Jeon M, et al.: Parasitic Network: Learning-based Network Downsizing of Very Deep Neural Networks for Computer Vision, 2018 International Conference on Control, Automation and Information Sciences, New York: Ieee, 2018: 378–383.
  • 11.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44. 10.1038/nature14539 WOS:000355286600030. [DOI] [PubMed] [Google Scholar]
  • 12.Huang J, Liu GX. Multi-color space threshold segmentation and self-learning k-NN algorithm for surge test EUT status identification. Front Mech Eng. 2016;11(3):311–5. 10.1007/s11465-016-0376-z WOS:000390958300010. [DOI] [Google Scholar]
  • 13.Liu GX, He BY, Liu SYA, Huang J. Chassis Assembly Detection and Identification Based on Deep Learning Component Instance Segmentation. Symmetry-Basel. 2019;11(8). 10.3390/sym11081001 WOS:000483559300011. [DOI] [Google Scholar]
  • 14.Guo YM, Liu Y, Oerlemans A, Lao SY, Wu S, Lew MS. Deep learning for visual understanding: A review. Neurocomputing. 2016;187:27–48. 10.1016/j.neucom.2015.09.116 WOS:000374606700005. [DOI] [Google Scholar]
  • 15.He K, Zhang X, Ren S, Sun J, editors. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition; 2016.
  • 16.Long J, Shelhamer E, Darrell T, Ieee. Fully Convolutional Networks for Semantic Segmentation. New York: Ieee; 2015. 3431–40 p. [DOI] [PubMed] [Google Scholar]
  • 17.Badrinarayanan V, Kendall A, Cipolla R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39(12):2481–95. 10.1109/TPAMI.2016.2644615 WOS:000414395400012. [DOI] [PubMed] [Google Scholar]
  • 18.Zhao H, Shi J, Qi X, Wang X, Jia J, editors. Pyramid scene parsing network. Proceedings of the IEEE conference on computer vision and pattern recognition; 2017.
  • 19.Chen L-C, Papandreou G, Kokkinos I, Murphy K, Yuille ALJapa. Semantic image segmentation with deep convolutional nets and fully connected CRFs. 2014. arXiv preprint arXiv:1412.7062. [DOI] [PubMed] [Google Scholar]
  • 20.Yu F, Koltun VJapa. Multi-scale context aggregation by dilated convolutions. 2015. arXiv preprint arXiv:1511.07122.
  • 21.Chen LC, Papandreou G, Kokkinos I, Murphy K, Yuille AL. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans Pattern Anal Mach Intell. 2018;40(4):834–48. 10.1109/TPAMI.2017.2699184 WOS:000426687100005. [DOI] [PubMed] [Google Scholar]
  • 22.Chen L-C, Papandreou G, Schroff F, Adam HJapa. Rethinking atrous convolution for semantic image segmentation. 2017. arXiv preprint arXiv:1706.05587. [Google Scholar]
  • 23.Lafferty J, McCallum A, Pereira FC. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. Proceedings of the Eighteenth International Conference on Machine Learning, 2001: 282–289.
  • 24.Huang J, Liu G, Wang B. Semantic Segmentation under a Complex Background for Machine Vision Detection Based on Modified UPerNet with Component Analysis Modules. Mathematical Problems in Engineering. 2020;2020:6903130 10.1155/2020/6903130 [DOI] [Google Scholar]
  • 25.Yu J, Kim D, Jeon M, editors. Parasitic Network: Learning-Based Network Downsizing of Very Deep Neural Networks for Computer Vision. 2018 International Conference on Control, Automation and Information Sciences (ICCAIS); 2018: IEEE.
  • 26.Lu QS, Liu CH, Jiang ZQ, Men AD, Yang B. G-CNN: Object Detection via Grid Convolutional Neural Network. IEEE Access. 2017;5:24023–31. 10.1109/access.2017.2770178 WOS:000417742800007. [DOI] [Google Scholar]
  • 27.Nowozin S, editor Optimal decisions from probabilistic models: the intersection-over-union case. Proceedings of the IEEE conference on computer vision and pattern recognition; 2014.
  • 28.Guo J, He H, He T, Lausen L, Li M, Lin HB, et al. GluonCV and GluonNLP: Deep Learning in Computer Vision and Natural Language Processing. J Mach Learn Res. 2020;21:7. WOS:000513691300023. [Google Scholar]
  • 29.He K, Zhang X, Ren S, Sun J. Identity mappings in deep residual networks. European conference on computer vision: Springer; 2016. p. 630–45.
  • 30.Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, et al., editors. The cityscapes dataset for semantic urban scene understanding. Proceedings of the IEEE conference on computer vision and pattern recognition; 2016.

Decision Letter 0

Jeonghwan Gwak

1 Dec 2020

PONE-D-20-27701

Fast Semantic Segmentation Method for Machine Vision Inspection Based on a Fewer-Parameters Atrous Convolution Neural Network

PLOS ONE

Dear Dr. Guixiong,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 15 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Jeonghwan Gwak, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please ensure that you refer to Figure 13 in your text as, if accepted, production will need this reference to link the reader to the figure.

Additional Editor Comments:

Please revise this manuscript to fully respond to reviewer 2's comments.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In this paper, the authors present fast semantic segmentation method for machine vision inspection based on a fewer parameters atrous convolution neural network. The article is well constructed, the experiments were well conducted, and analysis was well performed.

Reviewer #2: This work designed a fewer-parameters optimization mathematical model according to the atrous convolution architecture network to decrease segmentation time while maintaining the detection accuracy, although there are some significant improvements that the authors reflected previous reviewers, there are several minor aspects the authors should consider:

1. More recent works should be included, both CNN-based semantic segmentation and Atrous CNN-based works to enlighten your novelty in the related works section.

2. The authors merely presented their observations in the results section. However, to make the experimental results more convinced, please give more explanations on the results you obtained.

3. Please increase the figures' resolution if possible.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Feb 10;16(2):e0246093. doi: 10.1371/journal.pone.0246093.r003

Author response to Decision Letter 0


15 Dec 2020

Respones to Reviewers

Dear Editor,

This article have sent to Plos One as Semantic Segmentation Visual Detection Technology Based on Downsampling Porous Convolution Architecture Network Model Lightweight Optimization (ONE-D-20-15502), and receive a decision letter.

Thank you very much for your time and discussion on this manuscript. We also thank the Reviewers very much for their constructive comments. We have revised the manuscript accordingly. The replies are follows:

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming.

Reply: According to the editor’s suggestion, we edit our manuscript to meets PLOS ONE's style requirements, including those for file naming.

2. Please ensure that you refer to Figure 13 in your text as, if accepted, production will need this reference to link the reader to the figure.

Reply: According to the editor’s suggestion, we refer to Figure 13 in our text.

Additional Editor Comments: Please revise this manuscript to fully respond to reviewer 2's comments.

Reply: According to the editor’s suggestion, we revise this manuscript to fully respond to reviewer 2's comments.

Reviewer #1:

In this paper, the authors present fast semantic segmentation method for machine vision inspection based on a fewer parameters atrous convolution neural network. The article is well constructed, the experiments were well conducted, and analysis was well performed.

Reviewer #2:

This work designed a fewer-parameters optimization mathematical model according to the atrous convolution architecture network to decrease segmentation time while maintaining the detection accuracy, although there are some significant improvements that the authors reflected previous reviewers, there are several minor aspects the authors should consider:

1. More recent works should be included, both CNN-based semantic segmentation and Atrous CNN-based works to enlighten your novelty in the related works section.

Reply: According to the reviewer’s suggestion, we list more recent works in the related works section, both CNN-based semantic segmentation and Atrous CNN-based works that enlighten my novelty.

2. The authors merely presented their observations in the results section. However, to make the experimental results more convinced, please give more explanations on the results you obtained.

Reply: According to the reviewer’s suggestion, we give more explanations on the results that we obtained. Especially the CITYSCAPES pixel-level semantic labeling results and banknote anti-counterfeit segmentation results

3. Please increase the figures' resolution if possible.

Reply: According to the reviewer’s suggestion, we increase the figures' resolution to 600dpi, which is twice as high as the PLOS ONE requirements.

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool

Reply: As Request, We upload our figure files to the PACE digital diagnostic tool, and upload the PACE output figure files to PLOSONE.

Attachment

Submitted filename: Respones to Reviewers.docx

Decision Letter 1

Jeonghwan Gwak

14 Jan 2021

Fast Semantic Segmentation Method for Machine Vision Inspection Based on a Fewer-Parameters Atrous Convolution Neural Network

PONE-D-20-27701R1

Dear Dr. Guixiong,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jeonghwan Gwak, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors reflected all of the reviewers' comments. Especially, the authors improved their manuscript by adding more recent works in the related works section, and more explanations on the results. Overall, the manuscript is well constructed, the experiments were well conducted, and analysis was well performed. Also, I find the manuscript very clearly written. For these reasons, I am strongly inclined to recommend it for publication.

Reviewer #2: After reflecting previous reviewers' comments, some significant improvements have been made. In overall, the organization of the paper is well-constructed, the experiments are well-conducted, the results are well-analyzed and the manuscript is well-written.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Jeonghwan Gwak

28 Jan 2021

PONE-D-20-27701R1

Fast Semantic Segmentation Method for Machine Vision Inspection Based on a Fewer-Parameters Atrous Convolution Neural Network

Dear Dr. Guixiong:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jeonghwan Gwak

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: 0903Reply to Review Reports.pdf

    Attachment

    Submitted filename: Respones to Reviewers.docx

    Data Availability Statement

    The PASCAL VOC Dataset used to support the findings of this study is available at http://host.robots.ox.ac.uk/pascal/VOC/. The CITYSCAPES Dataset used to support the findings of this study is available at http://cityscapes-dataset.com. The DeepLab v3 pretrained model are the implement of GLUON-CV, and available at https://gluon-cv.mxnet.io/model_zoo/segmentation.html. We provide the DeepLab v3 with modified ResNet50, and a program (*.ipynb) to use it on different image of PASCAL VOC, CITYSCAPES or CNY. https://github.com/HJ0116/utilization_semantic_segmentation


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES