Skip to main content
Elsevier - PMC COVID-19 Collection logoLink to Elsevier - PMC COVID-19 Collection
. 2022 Jun 22;78:103933. doi: 10.1016/j.bspc.2022.103933

Two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm for lung parenchyma segmentation of COVID-19 CT image

Guowei Wang a,1, Shuli Guo a,1, Lina Han b,, Anil Baris Cekderi a
PMCID: PMC9217142  PMID: 35774106

Abstract

The lesions of COVID-19 CT image show various kinds of ground-glass opacity and consolidation, which are distributed in left lung, right lung or both lungs. The lung lobes are uneven and it have similar gray value to the surrounding arteries, veins, and bronchi. The lesions of COVID-19 have different sizes and shapes in different periods. Accurate segmentation of lung parenchyma in CT image is a key step in COVID-19 detection and diagnosis. Aiming at the unideal effect of traditional image segmentation methods on lung parenchyma segmentation in CT images, a lung parenchyma segmentation method based on two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm is proposed. Firstly, the optimal threshold method is used to realize the initial segmentation of the lung, so that the segmentation threshold can change adaptively according to the detailed information of lung lobes, trachea, bronchi and ground-glass opacity. Then the lung parenchyma is further processed to obtain the lung parenchyma template, and then the defective template is repaired combined with the improved Freeman chain code and Bezier curve. Finally, the lung parenchyma is extracted by multiplying the template with the lung CT image. The accuracy of lung parenchyma segmentation has been improved in the contrast clarity of CT image and the consistency of lung parenchyma regional features, with an average segmentation accuracy rate of 97.4%. The experimental results show that for COVID-19 and suspected cases, the method has an ideal segmentation effect, and it has good accuracy and robustness.

Keywords: Multi-threshold segmentation, COVID-19, Improved firefly algorithm, Spatial neighborhood information, Freeman chain code, Lung parenchyma

1. Introduction

1.1. Background problem domain

COVID-19 is an acute respiratory infectious disease caused by the type of new coronavirus infection, which is highly contagious [1], [2]. Until October 6, 2021, there have been more than 235.67 million confirmed cases of COVID-19 worldwide and 4.81 million cases died. With the rapid development of CT, CT images have been able to provide high-definition lung images with high contrast among lung tissues [3]. Therefore, high-resolution CT images are gradually used for the diagnosis of COVID-19 [4]. Computer aided detection (CAD) technology can effectively detect various ground-glass opacity from lung CT images, so as to lay a foundation for the early diagnosis and treatment of COVID-19 [5]. The main processes of CAD technology based on CT images include preprocessing, lung parenchyma segmentation, candidate lesion segmentation, feature extraction and optimal selection, disease classification, etc [6]. Among them, the most important step is lung parenchyma segmentation, and it is also a key link affecting the stability and accuracy of CAD. Rapid and accurate segmentation of lung parenchyma from CT images is the key to the diagnosis of COVID-19, which has very important practical significance and clinical value.

1.2. Review of literature

The main clinical symptoms of COVID-19 are fever, cough and fatigue, which may lead to acute respiratory distress syndrome (ARDS) [7], [8]. The main obstacle to controlling the spread of this disease is the lack of efficient detection methods. Reverse transcription polymerase chain reaction (RT-PCR) is the gold standard for the diagnosis of COVID-19, but it takes 4–6 h to obtain the results, and the RT-PCR test also has a high false negative rate [9]. It takes several times of RT-PCR test every few days to confirm the diagnostic results [10]. In addition, RT-PCR reagents are also insufficient in many severely epidemic areas [11]. In contrast, CT examination has been widely used in various hospitals. In clinical practice, CT is a more efficient and safer method to diagnose COVID-19 by combining clinical symptoms and signs [12]. At the same time, CT images can also be used in the tracking and prognosis of COVID-19. The size and shape of COVID-19 ground-glass opacity (GGO) are the main factors for COVID-19 tracking and prognosis. In the early stage of COVID-19, the density of GGO is low. The GGO is not obvious, and it is easy to cause the miss diagnosis. The CAD technology based on CT images can accurately classify COVID-19, and the real time observation of GGO changes can improve prognosis and reduce mortality. The GGO of COVID-19 involve both lungs. The manifestations of the lesions are diverse, and the CT images of GGO change rapidly. Therefore, CT diagnosis and timely treatment for COVID-19 patients are very important to improve tracking and prognosis of COVID-19. The lung image of early patients show multiple small patchy opacities. Advanced patients show multiple GGO and infiltration shadow of both lungs, and severe patients can have lung consolidation [13]. In conclusion, medical imaging plays a vital role in limiting virus transmission and diagnosing COVID-19.

In the medical image diagnosis of COVID-19, due to the need to screen and treat a large number of patients, in order to reduce the workload of doctors, artificial intelligence technology can be used to automatically diagnose the medical images and calculate the severity of infection [14], [15]. Jin [16] has designed an AI-assisted diagnosis system, which can achieve comparable performance to that of doctors, and the system has been deployed to many hospitals. In addition, self-diagnosis technology can also predict high-risk patients according to the clinical characteristics of COVID-19, which can help doctors find and treat the disease [17]. Wang [18] proposed the PatchShuffle convolutional neural network (PSCNN) and applied it for COVID-19 image classification to highlight the influence of neighboring pixels on the classification results of central pixels, so as to improve the classification accuracy of COVID-19. Zhang [19] divides the COVID-19 image into self-similar blocks and non-self-similar blocks by using the Pseudo Zernike moment and deep stacked sparse autoencoder. The Pseudo Zernike moments of the self-similar blocks are calculated to select the most robust moment. The CT image of COVID-19 is diagnosed by quantizing and modulating the amplitude of the pseudo Zernike moments.

In the lung CT image, the structure of each tissue is complex and the shape is different. The difference of gray contrast between the lung and the surrounding tissues (muscles, blood vessels) is low. These factors increase the difficulty of lung parenchymal segmentation. Threshold segmentation is a method of image segmentation, which segments the target and background in the image by selecting an appropriate threshold. The threshold segmentation method divides the original CT image pixel into several subsets by setting a threshold to segment the lung parenchyma. It mainly uses the difference in grayscale between the lung parenchyma to be extracted from the CT image and the lung organs to perform image segmentation. The traditional segmentation threshold method of lung images with a small time complexity, but its stability is not high [20]. When the threshold is not selected properly, it is easy to ignore the edge lesions and the surrounding tissues. The gray contrast is not obvious and the region of interest may be over segmented. The appropriate threshold method can achieve better segmentation, and it segments the CT image into lung parenchyma target region and background region. When the image is complex or there are many target objects [21], [22], the multi-threshold segmentation is needed. It aims to effectively segment background and multiple targets, which is an extension of single threshold segmentation.

The traditional threshold segmentation methods find the appropriate threshold by setting specific criteria, and the operation speed is slow, such as Otsu method [23], minimum error method [24], image entropy method [25] and so on. Many scholars have proposed improved algorithms to increase the speed of segmentation. Raja [26] used particle swarm optimization (PSO) with improved inertia factors to optimize the Otsu threshold segmentation method. Rajinikanth [27] proposed a fast threshold segmentation method based on Otsu criterion and line intercept histogram. Wang [28] proposed an improved adaptive genetic algorithm, combined with the two-dimensional fisher image segmentation evaluation function for global optimization, in order to improve the speed of the segmentation threshold. In recent years, image entropy threshold segmentation has been widely used in image segmentation. Zhou [29] used the maximum entropy of one-dimensional grayscale to segment gray image effectively, but the algorithm is sensitive to noise. In order to overcome the noise sensitivity, Estanon [30] proposed a threshold segmentation method of two-dimensional entropy. This method makes full use of the gray value and local spatial information of the image, and it can achieve good segmentation results. But this method still has the problem of time-consuming.

Most of the threshold segmentation methods are time-consuming, while the multi-threshold segmentation method has a larger amount of calculation, so its calculation time is longer and it occupies more memory. Aiming at this problem, many scholars combine the multi-threshold segmentation with the optimization algorithm. Chakraborty [31] used PSO algorithm, Darwin particle swarm optimization (DPSO) and fractional order Darwin particle swarm optimization (FODPSO) to find the optimal threshold under the criterion of maximum variance between classes respectively. Horng [32] proposed to find the optimal threshold under the minimum cross entropy by using the honey-bee mating optimization (ABCA) algorithm. Ozyurt [33] combined genetic algorithm (GA) and improved maximum fuzzy entropy for multi-threshold segmentation. These methods improve the speed of finding multi-threshold, but PSO, GA and ABCA algorithm are easy to fall into local extremum, which makes the optimization results inaccurate.

Hrosik [34] proposed a firefly algorithm (FA) for multi-mode optimization. The algorithm is optimized by the phototaxis between fireflies, and it has the ability of fast global search. Wu [35] solved multiple constrained optimization problems by using firefly algorithm. Horng [36] combined the firefly algorithm and the minimum cross entropy criterion for image multi-threshold segmentation, and it achieved a good result. In this paper, the firefly algorithm is combined with two-dimensional entropy multi-threshold segmentation method to segment multi-target images.

There are some state-of-the-art methods about image segmentation algorithm reported. Alqazzaz [37] proposed an improved SegNet model for image segmentation. Through the global gray information to guide the curve to approach the target edge, it has stronger segmentation adaptability for the image with uneven foreground gray level. However, this model does not abstract the higher-level semantic features for image segmentation. Therefore, when the target size is small, the background is complex, and the features are not significant, the image segmentation effect is still not good. Ghaffari [38] proposed a fast and robust fuzzy c-means clustering algorithm (FRFCM) based on the morphological reconstruction and membership filtering. This algorithm effectively uses the local spatial information of the image and it replaces the traditional distance calculation with the membership filtering to ensure the superiority of anti-noise performance. But the clustering result is too smooth, and too much detailed information will be lost. Li [39] proposed a contextualized convolutional neural network (Co-CNN) structure. This method integrates the multi-level image content into a unified network, and the weighted multiscale features are trained at each pixel position. However, this method is realized by introducing different scales of input, which cannot realize the human-like attention mechanism, and the algorithm runs for a long time. Basukala [40] proposed a dual tree-complex wavelet transform (DT-CWT) image segmentation algorithm. MRF sequences are constructed in wavelet domain multi-resolution, which are nested with each other, and this method realizes the effective image segmentation by recursive operation between adjacent scales in the wavelet domain. This algorithm not only represents the image structure information, but also better describes the instability of the image. However, the wavelet transform lacks translation invariance, and it can only model the feature information of the image in the horizontal, vertical and 45 directions. Therefore, the continuity of image segmentation based on wavelet is not ideal in obtaining the target edge, and there is a certain “hole” phenomenon in the detection result. Yan [41] proposed a GAN-Unet image segmentation method. This model consists of two parts including 2D and 3D Unet. The 2D part extracts the features in the slice, and the 3D part integrates the information of the slice. The network can be used for organ and lesion segmentation. However, because the model uses 3D convolutional network, it becomes difficult to train and it requires higher demand for hardware.

1.3. Highlights

The highlights of this work can be summarized as the following: (i) a multi-threshold segmentation method based on two-dimensional reciprocal cross entropy is proposed to make up for the problem of undefined and zero value of Shannon cross entropy due to logarithm operation; (ii) aiming at the problem the traditional histogram only considers the background and the target area, the edge and the noise information are ignored. This paper improves the segmentation of the binary histogram that is divided the four-region histogram into two-region of the target and the background, and the gray information is fully considered; (iii) traditional threshold segmentation algorithms only use the value of gray level in the image without considering the spatial neighborhood information of pixels, which greatly increases the ratio of misclassification between the target and the background of the gray image. Aiming at this problem, a weighted fuzzy threshold segmentation based on the spatial neighborhood information is proposed; (iv) aiming at the problem that the standard firefly algorithm is easy to fall into local optimization, this paper improves the attraction term by using exponential distribution in the position movement and the exponential decreasing inertia weight to enhance the global detection ability. The step size monotone decreasing mode is used to improve the random item to enhance the local mining ability in the later stage of the optimization; (v) the lung parenchyma template is repaired by the proposed eight-direction Freeman chain code combined with quadratic Bezier curve can make up for the under-segmentation problem when repairing large-area defects with straight lines. It has flexible adjustment to better meet the characteristics of lung parenchyma edge, which can achieve a good repair effect.

1.4. Paper structure

The rest of this paper is organized as follows. Section 2 describes the two-dimensional reciprocal cross entropy multi-threshold image segmentation based on the two-region histogram and spatial neighborhood information. In Section 3, we propose the improved firefly algorithm to optimize the parameters of the two-dimensional reciprocal cross entropy. Section 4 describes the refinement segmentation of lung parenchyma. The defective template is repaired combined with the improved Freeman chain code and Bezier curve, and then the lung parenchyma is extracted by multiplying the template with the lung CT image. In Section 5, we show the simulation experiments for the different kinds of COVID-19 CT images. The paper is structured as shown in Fig. 1.

Fig. 1.

Fig. 1

Paper structure.

2. Two-dimensional reciprocal cross entropy image segmentation

2.1. Two-dimensional reciprocal cross entropy single-threshold image segmentation

The size of the gray image I is set to M×N. I(x,y) represents the gray value of the pixel (x,y) in the image I. 1xM, 1yN, and the gray level of the image is L. The average gray value of the n×n neighborhood corresponding to each pixel (x,y) of the image I is defined as G(x,y), and its mathematical expression is: G(x,y)=1n2i=n/2n/2j=n/2n/2I(x+i,y+j).

The frequency of simultaneous occurrence of the pixel gray value i at a certain position in the image I(x,y) and the pixel gray value j at the corresponding position of the neighborhood average gray image G(x,y) is set as rij. The corresponding joint probability density is defined as pij=rij/(M×N)(i,j=0,1,,L1). Where: i=0L1j=0L1pij=1. pij constitutes a two-dimensional histogram of the image, as shown in Fig. 2. Zone A and zone B represent the background and goal respectively; Zone C and zone D represent the edge and noise respectively (zone C and zone D are usually ignored when calculating the reciprocal cross entropy). The probabilities of zone A and zone B are: PA=i=0sj=0tpij, PB=i=s+1L1j=t+1L1pij.

Fig. 2.

Fig. 2

Two-dimensional histogram.

The normalized probabilities corresponding to the gray level in zone A and zone B are: PijA=piji=0sj=0tpij, PijB=piji=s+1L1j=t+1L1pij.

The reciprocal cross entropy of the image corresponding to the threshold (s,t) is defined as:

H(s,t)=H(A)+H(B)=i=0sj=0tijpij1+ijPijAi=s+1L1j=t+1L1ijpij1+ijPijB (1)

When H(s,t) takes the minimum value, the best single threshold of the two-dimensional reciprocal cross entropy can be obtained: (s˙,t˙)=argmin0sL1min0tL1H(s,t).

2.2. Two-dimensional reciprocal cross entropy multi-threshold image segmentation based on the improved histogram

The traditional histogram only considers the background and the target area, but the edge and the noise information in area C and D are ignored, resulting in poor effect of the image segmentation. When calculating the optimal threshold, if the area C and D are ignored, the points with large difference between the pixel gray level and the average gray level of the neighborhood are not fully considered. These points may exist in area C and D, and the calculated optimal threshold will be biased and the segmentation results are inaccurate. In order to improve the quality of segmentation, this paper improves the division of the binary histogram, as shown in Fig. 3.

Fig. 3.

Fig. 3

Segmentation of improved histogram.

In Fig. 3, C, Co and Cb represent the whole area, target and background area respectively. The equations are:

C={(i,j)|i=0,1,,L1;j=0,1,,L1}={(x,y)|I(x,y)=0,1,,L1;G(x,y)=0,1,,L1} (2)
Co={(i,j)|i=0,1,,t;j=0,1,,s}={(x,y)|I(x,y)=0,1,,t;G(x,y)=0,1,,s} (3)
Cb=CCo (4)

Where, I(x,y) represents the gray value of the pixel (x,y) in the image I; G(x,y) represents the average gray value of the n×n neighborhood corresponding to each pixel (x,y) of the image I. The threshold is used to change the four-region segmentation of the histogram into two-region segmentation of the target and the background, and the gray information of the target and the background is fully considered, then the normalized probabilities corresponding to the gray level in zone o and zone b are: Pijo=piji=0tj=0spij, Pijb=piji=t+1L1j=s+1L1pij.

The two-dimensional reciprocal cross target entropy and background entropy are respectively: H(o)=i=0tj=0sijpij1+ijPijo, H(b)=i=t+1L1j=s+1L1ijpij1+ijPijB. The total entropy of two-dimensional reciprocal cross is: H(t,s)=H(o)+H(b).

The lesions of COVID-19 CT image showed various ground-glass opacity and consolidation. The lung lobe is uneven and it has similar gray value to the surrounding arteries, veins and bronchus. Aiming at the problems of low contrast and fuzzy boundary in the traditional two-dimensional single-threshold lung image segmentation algorithm, it is proposed to extend the two-dimensional reciprocal cross entropy single-threshold to multi-threshold segmentation. The image I is segmented by n1 gray levels, and the two-dimensional reciprocal cross entropy single-threshold segmentation is extended to multi-threshold. The values of the threshold are (t1, s1)(t2, s2)...(tn1, sn1).

As shown in Fig. 4, the image P is segmented into n classes {C1,C2,,Cn}. The normalized probability corresponding to the gray level Ck(k=1,2,,n) is:

PijCk=piji=tk1+1tkj=sk1+1skpij (5)

Fig. 4.

Fig. 4

Segmentation of two-dimensional multi-threshold histogram.

Assuming that {i1, i2, …, im+1} is a subset of {1,2,,n}, and the total entropy of the two-dimensional reciprocal cross is:

H(t1,s1,t2,s2,,tn1,sn1)=H(C1)+H(C2)++H(Cn)=i=0t1j=0s1ijpij1+ijPijC1i=t1+1t2j=s1+1s2ijpij1+ijPijC2i=tk+1L1j=sk+1L1ijpij1+ijPijCn (6)

2.3. Calculation of spatial neighborhood information

The traditional image threshold segmentation method only uses the gray value information in the image pixels but ignores the influence of the spatial information between the pixels, which makes the segmentation not very effect. In this paper, a threshold segmentation algorithm combined with fuzzy theory is proposed. The algorithm comprehensively uses the gray value and the spatial neighborhood information of image pixels to design the membership function and the evaluation standard of fuzzy distance to solve the optimal threshold solution.

2.3.1. Calculation of similarity between pixels in spatial neighborhood

This function is to calculate the similarity between the pixels in the neighborhood of the window and the central pixel. For x(m,n,fmn)W(x,y), there is the following equation:

w(I(x,y),x(m,n,fmn))=1|I(x,y)x(m,n,fmn)|σ=1t=0nσ(t)|I(x,y)(t)x(m,n,fmn)(t)|=1(0.5×d1(I(x,y),x(m,n,fmn))×h+0.5×d2(I(x,y),x(m,n,fmn))×c) (7)

Where, I(x,y) means that the value of pixel in the position (x,y). W(x,y) represents the range of window centered on pixel I(x,y), this paper selects eight neighborhoods for the range of window. x(m,n,fmn) represents the pixels within the window. w(I(x,y), x(m,n,fmn)) represents the similarity weight between the pixels in the neighborhood of the window and the center pixel. σ(t) is the weight in the tth similarity measure. In this paper, two measures are taken: the geometric distance between the central pixel and the neighborhood pixels, the Euclidean distance of the gray value. n is the number of the similarity measure; |I(x,y)(t)x(m,n,fmn)(t)| represents the similarity degree of the t-th similarity measure between pixels; h is the maximum geometric distance from the central pixel in the neighborhood; c is the normalization factor.

2.3.2. Calculation of fuzzy membership degree of spatial neighborhood information

This function is to calculate the membership degree of the spatial neighborhood information: in region Ci, the similarity of the pixels is calculated whose gray value is less than or equal to (ti,si) in the neighborhood of the center pixel as shown in Eq. (7), then sum and its calculation is shown in Eq. (8). Similarly, the similarity of the pixels is calculated whose gray value in the neighborhood of the window is greater than (ti,si) as shown in Eq. (9). Finally, the spatial neighborhood information is transformed into the range of (0,1) by normalization processed according to Eq. (10) and Eq. (11).

μnsum1=s=0n1w(I(x,y),x(m,n,fmn)) (8)
μnsum2=s=0n2w(I(x,y),x(m,n,fmn)) (9)
μn1=μnsum1×n11+μnsum1×n1+μnsum2×n2 (10)
μn2=μnsum2×n21+μnsum1×n1+μnsum2×n2 (11)

Where, n1 and n2 are the number of pixels in the window neighborhood under the conditions of Eq. (8) and Eq. (9) respectively.

2.3.3. Calculation of fuzzy membership function of gray value information

In the region Ci, the gray histogram of the image is calculated, and the whole gradient image is divided into two categories by using the threshold (ti,si). The value of threshold less than or equal to (ti,si) is called background, and the value of threshold greater than (ti,si) is called foreground. On this basis, the calculation of the mean value μ0 for the background class can be obtained:

μ0=i=ti1+1tij=si1+1siI(x,y)×hist(I(x,y))i=ti1+1tij=si1+1sihist(I(x,y)) (12)

Similarity, the calculation of the mean value μ1 for the foreground class can be obtained:

μ1=i=ti+1ti+1j=si+1si+1I(x,y)×hist(I(x,y))i=ti+1ti+1j=si+1si+1hist(I(x,y)) (13)

Where, I(x,y) in Eq. (12) and Eq. (13) means that the value of pixel in the position (i,j) is fij. hist(I(x,y)) is the number of pixels whose value of pixel at (i,j) is fij.

The fuzzy membership of the gray information of the image pixels is calculated: the maximum and minimum gray values are counted by scanning the gray information of each pixel in the image. In the region Ci, assuming that the threshold is (ti,si), the mean value is solved by Eq. (12) and Eq. (13), and then the gray information value of each pixel is taken out. The fuzzy membership degree of the gray information of the corresponding image pixels is calculated by Eq. (14) and Eq. (15), and the form is following:

μg1(I(x,y))=1|I(x,y)μ0|I(x,y)fmaxI(x,y)fmin (14)
μg2(I(x,y))=1|I(x,y)μ1|I(x,y)fmaxI(x,y)fmin (15)

Where, I(x,y)fmax is the maximum gray value of the image pixel; I(x,y)fmin is the minimum gray value of the image pixel; the maximum gray value and the minimum gray value are not equal; 0μgi(I(x,y))1,i=1,2.

2.3.4. The total membership function of pixels

In region Ci, the sum of the total membership degree of the image pixels greater than the threshold (ti,si) and less than the threshold (ti,si) is calculated respectively. Firstly, each pixel in the image is scanned, then the gray information and the spatial neighborhood information of the pixel are calculated respectively. The membership degree of the pixel that is less than the threshold (ti,si) is counted in Eq. (16). The membership degree of the pixel that is greater than the threshold (ti,si) is counted in Eq. (17). Finally, the values of the two parts are compared. If the result of Eq. (16) is greater than or equal to the result of Eq. (17), the pixel belongs to the class less than or equal to (ti,si), otherwise it belongs to class greater than (ti,si). The calculation is as following:

μ(I(x,y))=(μn1+μg1(I(x,y)))/2 (16)
μ(I(x,y))=(μn2+μg2(I(x,y)))/2 (17)

2.3.5. Calculation of the optimal threshold

Fuzzy distance is defined: A is the discrete fuzzy set defined on interval [μmin, μmax], and μA(u) is its membership function, then defining:

ϕ(A)=i=0nμiμmaxμmin (18)
ϕ(A)=ϕ(A)ϕ(U) (19)

Where, U is a finite element fuzzy set on [μmin, μmax] and μ(u)=1uU, namely:

ϕ(U)=f=fminfmaxhist(I(x,y)(f)) (20)

Where, hist(x(i,j,fij(f))) represents the number of pixels whose pixel value is f in the image. The distance between fuzzy sets A and B is defined as:

d(A,B)=|ϕ(A)ϕ(B)|1+ϕ(A)+ϕ(B) (21)

ϕ(A) and ϕ(B) can be regarded as the quotient of the membership degree and its membership span of all elements in fuzzy sets A and B. When ϕ(A) and ϕ(B) are constant, d(A,B) increases with the increase of the absolute value of the difference between ϕ(A) and ϕ(B). ϕ(A) and ϕ(B) are approximately equal to the average value of the membership degrees of all elements in the set A and B. When the absolute value of the difference between ϕ(A) and ϕ(B) is the same, the larger the ϕ(A) and ϕ(B) are, and the smaller the d(A,B) is. It can be seen that the greater the d(A,B), the greater the distance between the two sets is. The greater the value of distance, the clearer the difference between the image target and the background is, and the better the image segmentation effect is. Therefore, maximizing the distance between the two sets means minimizing the probability of misclassification between the target and the background.

The total entropy of the two-dimensional reciprocal cross based on the two-region histogram and the spatial neighborhood information is:

H(t1,s1,t2,s2,,tn1,sn1)=d(C1,C2)H(C1)+d(C1,C2)H(C2)d(C2,C3)+d(C2,C3)H(C3)d(C3,C4)+...+d(Cn1,Cn)H(Cn)=|ϕ(C1)ϕ(C2)|1+ϕ(C1)+ϕ(C2)i=0t1j=0s1ijpij1+ijPijC1|ϕ(C1)ϕ(C2)|1+ϕ(C1)+ϕ(C2)i=t1+1t2j=s1+1s2ijpij1+ijPijC2|ϕ(C2)ϕ(C3)|1+ϕ(C2)+ϕ(C3)|ϕ(C2)ϕ(C3)|1+ϕ(C2)+ϕ(C3)i=t2+1t3j=s2+1s3ijpij1+ijPijC3|ϕ(C3)ϕ(C4)|1+ϕ(C3)+ϕ(C4)......|ϕ(Cn1)ϕ(Cn)|1+ϕ(Cn1)+ϕ(Cn)i=tk+1L1j=sk+1L1ijpij1+ijPijCn (22)

The criterion of two-dimensional reciprocal cross entropy multi-threshold image segmentation is to find the threshold (t1˙,s1˙)(t2˙,s2˙)...(t˙n1,s˙n1) to minimize the two dimensional reciprocal cross entropy H(t1,s1,t2,s2,,tn1,sn1).

3. Two-dimensional reciprocal cross entropy multi-threshold image segmentation combined with improved firefly algorithm

In order to solve the problems of large amount of calculation and time-consuming in the multi-threshold image segmentation method, by improving the firefly algorithm, this paper proposes the two- dimensional reciprocal cross entropy multi-threshold image segmentation combined with the improved firefly algorithm, so as to transform the calculation of the multi-threshold segmentation into the optimization of the two-dimensional reciprocal cross entropy function H(t1,s1,t2,s2,,tn1,sn1) by the firefly algorithm.

The firefly algorithm is a bionic optimization method whose principle is to simulate individual firefly individuals in nature with points in the search space [42]. By using the phototaxis of fireflies in the process of optimization, the objective function of solving the problem is transformed into finding the firefly with the largest brightness. In each iteration, the position of the firefly is updated by attracted and moving to find the firefly with the greatest brightness [43]. Therefore, the firefly algorithm can carry out the global optimization quickly. The formula for the movement of the firefly i to the brighter firefly j is defined as:

Xi(t+1)=Xi(t)+β(γ)×(Xj(t)Xi(t))+α×(rand0.5) (23)

Where, Xi and Xj represent the position of firefly i and firefly j respectively; β(γ)=β0eγγij2 is the attraction of firefly j to firefly i; γij is the distance between fireflies i and j, γij=XiXj; β0 is the degree of attraction when γij is zero; γ is the absorption coefficient of light intensity; α is the step factor and it is a constant within [0,1]; rand is a random factor submitted to uniform distribution on [0,1]; α×(rand0.5) is the perturbation term, which is used to avoid falling into local optimization too early.

3.1. Limitation analysis of traditional firefly algorithm

The firefly algorithm mainly relies on the best individual in the field to guide the optimization. Once the field is an empty set, the algorithm will stagnate, which greatly reduces the optimization efficiency [44]. In addition, the location update formula of the algorithm is too simple and the search step is fixed, so that the algorithm cannot quickly optimize the optimal solution [45]. The lesions of COVID-19 CT image can be seen in left lung, right lung or both lungs. Aiming at the problems of complex structure and uneven gray distribution of various tissues in lung CT image, it is difficult to accurately segment and extract lung parenchyma and lesions. The exponentially decreasing inertia weight and random mechanism are proposed to make the segmentation threshold change adaptively according to the detailed information of lung lobes, trachea, bronchi and carina. The attraction term of firefly algorithm is matched and switched by combining exponential distribution and weber distribution, so as to obtain more accurate contour boundaries of each target with relatively fewer iterations.

3.2. Improved firefly algorithm based on the best position of the population

The principle of the improved firefly algorithm based on the optimal position of the population is shown in Fig. 5.

Fig. 5.

Fig. 5

Improved firefly algorithm based on the best position of population.

The position update formula of the firefly i under the influence by the optimal position of the population is:

Xi(t+1)=Xi(t)+β0×eγRij2×(Xj(t)Xi(t))+β0×eγRi,best2×(Xbest(t)Xi(t))+α×(rand0.5) (24)

Where, β0×eγRi,best2×(Xbest(t)Xi(t)) represents the update position of firefly i under the influence by the optimal position of the population.

3.3. Improved firefly algorithm based on random mechanism

The balance between detection and search ability is the core issue of the firefly algorithm. The detection ability can be improved by the random mechanism, which can make the algorithm jump out of the local optimization with a greater probability to realize the global search. The improvement of search ability depends on local information, such as gradient, mode and the historical information in the process of the algorithm.

Aiming at the problem of traditional firefly algorithm that is easy to fall into local optimality, the attraction is improved by using exponential distribution in the position movement formula, so as to enhance the global detection. In addition, the random item is improved by the step-size monotonous decreasing mode to enhance the local mining ability in the later stage of the algorithm. In this paper, random factor is added to disturb the attraction coefficients and it enhances the ability to jump out of the local optima. The expression of β is:

β=β0eγγij2Ri (25)

Where, Ri is the exponential distribution. If the step factor α takes a fixed value, the detection and search ability of the algorithm cannot be effectively adjusted. In this paper, α is decremented so that the random function has a large random step at the beginning of iteration, and the algorithm can search better in the global scope. The step size is gradually reduced in the later stage of the iteration, so that the algorithm can refine the search in the local scope. The expression of α is:

α=α0δtS,0<δ<1 (26)

Where, α0 is the initial random step size, δ is the cooling coefficient, 0¡δ¡1, S is the step factor. In the early stage of the algorithm, the larger step search is required to traverse a wide range of feasible solutions. In the later stage of the algorithm, the position function has slowly converged near the optimal solution. At this time, a small step is required to search. According to this improved idea, it can be found that in the tangent function tan(x), when the independent variable x is 0.785, the value of function is 1, which decreases with the decrease of the independent variable, and the reduction rate also decreases gradually. Therefore, the tangent function is introduced into the step factor, and the expression is:

S=tan(0.785×tmaxttmax)s0 (27)

Where, tmax is the maximum number of iterations; t is the maximum number of iterations; s0 is the fixed step parameter.

3.4. Exponentially decreasing inertia weight

The inertia weight w(t) is used to control the influence on the current position by the previous position of the firefly, which adjusts the global optimization. At the beginning, the inertia weight is large, which makes the relative attraction small. The global optimization is enhanced and the local search ability is weakened. On the contrary, with the increase of the iteration, the inertia weight gradually decreases, and the local search ability is enhanced, so as to avoid repeated oscillation on the extreme point due to the excessive movement of the firefly. According to the idea of linear decline, combined with the random and exponential inertia weight in particle swarm optimization, an improved exponential decline inertia weight is proposed, and the random disturbance term is added for adjustment. The weight formula is expressed as:

w(t)=wmin(wmaxwmin)1tT+randwmin (28)

Where, t is the current number of iteration, and wmax=0.8, wmin=0.2, T=150. rand is the random number between [0,1] to ensure w(t)(0,1). In summary, the position movement formula of the improved firefly algorithm is:

Xi(t+1)=w(t)Xi(t)+β0eγRij2Ri(Xj(t)Xi(t))+β0eγRi,best2Ri(Xbest(t)Xi(t))+α0δtS(rand0.5) (29)

Where: Xi(t+1) is the location of the firefly after updating; Xbest(t) is the optimal individual of the current population.

3.5. Time computational complexity of improved firefly algorithm

For the time computational complexity, O(f) is set as the time computational complexity in each iteration of each firefly, then it can be seen that the time computational complexity of the traditional firefly algorithm is O(TNf). Where, T is the maximum number of iteration and N is the maximum population size. For the traditional firefly algorithm, the value of the inertia weight in each iteration remains unchanged, so there is O(f1)=O(f2)==O(fT)=O(f). For the proposed firefly algorithm, the value of the inertia weight gradually decreases with the increase of iteration, so there is O(f1)O(f2)O(fT). It can be seen from the above analysis that compared with the traditional firefly algorithm, the proposed firefly algorithm greatly reduces the time computational complexity.

3.6. Steps and flow chart of multi-threshold image segmentation method

The two-dimensional reciprocal cross entropy H(t1,s1,t2,s2,,tn1,sn1) is set as the objective function of the firefly algorithm, and the optimization result is the firefly position X(t1˙,s1˙,t2˙,s2˙,,t˙n1,s˙n1) with the highest brightness. Where: k is the number of threshold, (t1˙,s1˙)(t2˙,s2˙)...(t˙n1,s˙n1) is the required threshold.

The steps of the two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm are as following:

(1) The parameters of firefly algorithm are initialized, including the number of fireflies n, the initial position Xi(ti1,si1,ti2,si2,,ti(n1),si(n1)), the initial attraction β0, the light intensity absorption γ, the step factor α0, and the maximum number of iteration T, the fixed step parameter s0. We use the horizontal test [46] to obtain the combination with the most occurrences of the optimal solution, then the number of fireflies n and the maximum number of iteration T are fixed. The enumeration method [47] is used to obtain the range of a series optimal value combinations for the initial attraction β0[0,2], the light intensity absorption γ[0.01,10], the step factor α0[0,1], and the fixed step parameter s0[0,0.8]. Finally, we analyze the results to obtain the optimal parameters of the improved firefly algorithm. β0=1, α0=0.5, n=100, T=150, γ=1, s0=0.4;

(2) The two-dimensional reciprocal cross entropy function Hi(ti1,si1,ti2,si2,,ti(n1),si(n1)) is calculated. We sort the value of objective function to obtain the position of the firefly with the largest brightness;

(3) We judge whether the maximum number of iteration T is reached. If it is reached, turn to Step.(4), otherwise turn to Step.(5);

(4) The position of the firefly with the highest brightness is output, and the calculated is used as the threshold to segment the image;

(5) The position of the firefly is updated according to Eq. (29), randomly disturb the fireflies in the best position, increase the search times by 1, and turn to Step.(2) for the next search.

The flow chart is shown in Fig. 6. Taking early COVID-19 as an example, the CT image after multi-threshold image segmentation is shown in Fig. 11(a).

Fig. 6.

Fig. 6

The flow chart of two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm for image segmentation.

Fig. 11.

Fig. 11

CT image of early COVID-19 segmented by TDRMFA: (a)binarized segmentation; (b)removal of trachea and bronchus; (c)removal of bed board and elimination of holes; (d)separation of the left and right lungs; (e)edge repair; (f)lung parenchyma segmentation. CT image of early COVID-19 segmented by SFTM: (a1)binarized segmentation; (b1)removal of trachea and bronchus; (c1) removal of bed board and elimination of holes; (d1)separation of the left and right lungs; (e1)edge repair; (f1)lung parenchyma segmentation. CT image of early COVID-19 segmented by PTEM: (a2) binarized segmentation; (b2)removal of trachea and bronchus; (c2)removal of bed board and elimination of holes; (d2)separation of the left and right lungs; (e2)edge repair; (f1)lung parenchyma segmentation. CT image of early COVID-19 segmented by PDEE: (a3)binarized segmentation; (b3) removal of trachea and bronchus; (c3)removal of bed board and elimination of holes; (d3)separation of the left and right lungs; (e3)edge repair; (f3)lung parenchyma segmentation.

4. Refinement segmentation of lung parenchyma

4.1. Removal of trachea and bronchus based on morphology

The lung parenchyma is more obvious after the initial segmentation of CT image through the above optimal threshold iterative method, but the presence of blood vessels and trachea in the image has a great impact on the accurate segmentation of the lung area. Therefore, the processed binary image is separated from the trachea adhering to the lung parenchyma by mathematical morphology [48]. Then combined with the hole filling algorithm to remove the lung trachea tree and small-area lung parenchyma holes, the strel function expansion corrosion is used to remove the boundary irrelevant to the lung parenchyma in the image and smooth the edge of the lung parenchyma, in order to obtain a complete lung parenchyma template. Taking early COVID-19 as an example, the CT image after removal of the trachea and bronchi is shown in Fig. 11(b).

4.2. Removal of bed board and elimination of holes based on closed operation

It can be seen from the binary image of lung parenchyma that there are a large number of holes inside the lung parenchyma and bed shadow outside the lung parenchyma. Moreover, due to the existence of X-ray pulse interference, there are many isolated noises in CT images. Therefore, in order to improve the quality of image, the holes inside the lung parenchyma are filled, the boundary of the lung parenchyma is smoothed, and the shape and size of the lung parenchyma remain unchanged. This paper proposes the closed operation of binary morphology to further process the lung parenchyma image obtained in Section 4.1.

The closed operation in binary morphology is defined as a mathematical morphological operator that uses the same structure element to perform the dilation and erosion operations on binary images successively. Where, the structural elements are the basic operators in binary morphology. Through the specific logical operation with the surrounding area of each pixel in the lung parenchyma image, the new value of each pixel is obtained. The structural elements include circle, rectangle, diamond, straight line and other shapes. In view of the advantages of isotropy and strong operability for the square structural elements, the closed operation is carried out by using the square structural elements.

According to the sequence of the closed operation, the lung parenchyma image is expanded firstly. The expansion operation can fill the holes in the lung parenchyma and smooth the boundary of the lung parenchyma. Assuming that the lung parenchyma image is F and the structural element is J, the structural element J is moved sequentially on the image plane. When passing through each pixel, J[s] has the following three states: (1)J[s1]F; (2)J[s2]Fc; (3)J[s3]F,J[s3]Fc; as shown in Fig. 7.

Fig. 7.

Fig. 7

Three possible states of J[s].

The expansion operation corresponds to the third state above. J[s] is related to part F, and each pixel in F is expanded into structural element J to obtain a new image, which can be described as:

FJ={s|J[s]F} (30)

The erosion operation can remove the isolated noise and restore the size of lung parenchyma at the same time. The erosion operation corresponds to the first state above, and J[S] is most correlated to F. The image after the erosion operation can be described as:

FJ={s|J[s]F} (31)

In summary, the image obtained after the closed operation is:

CLOSE(F)=(FJ)J (32)

Taking early COVID-19 as an example, the CT image after removal of bed board and elimination of holes is shown in Fig. 11(c).

4.3. Separation of the left and right lungs

Due to the complex structure of the two lung regions, and the similar tissue gray levels or the influence of lung disease, it is difficult to define the edges of the two lung regions, which will interfere with the detection of lung disease in the later stage. Therefore, it is necessary to separate the left and right lungs in the process of lung parenchyma segmentation. In this paper, the two lung regions are separated by combining the grayscale integral projection and scanning algorithm [49]. Firstly, whether there is adhesion between the left and right lungs is judged, and each connected domain in the image is marked and then its size in descending order is arranged. The ratio of the minimum and maximum area of the pixel column coordinate connected domain is recorded as S1, and the ratio of the minimum and maximum area of the pixel row coordinate connected domain is recorded as S2. When S1>0.5 and S2>0.5, it is judged that the left and right lungs are separated, otherwise, the left and right lungs are adhered. Secondly, the position of the connection between the two lungs is determined according to the gray scale integral projection feature of the image. Finally, the two lung regions are made independent by the rule-based row and column scanning. Taking early COVID-19 as an example, the result of lung parenchyma separation is shown in Fig. 11(d).

4.4. The edge repair template of lung parenchyma

If the edge of the processed lung parenchyma template is complete and smooth, it indicates that there is no nodule at the edge, which does not affect the integrity of the lung parenchyma and it is directly saved as a template. If the edge of the template is rough and concave at this time, there may be neglected fuzzy problems such as edema, inflammation or nodules. At this time, the template needs to be repaired.

The convex hull algorithm [50] fails to achieve the desired ideal effect when inpainting the lung parenchyma template of large area lesions. The rolling ball method cannot accurately determine the size of the defect when repairing the edge of the lung parenchyma template, and it is difficult to accurately set the radius of the sphere. If the radius is not selected properly, it is easy to have the problem of under segmentation or over segmentation.

The Freeman chain code can provide sufficient information about the target image and it is not affected by image noise. It represents the boundary of the image by line segments of unit length and preset direction. In this paper, the Freeman chain code and the difference of chain code are used to describe the concavity and convexity of the boundary to find the defect position of the template. Firstly, the starting position of the chain code needs to be determined and marked. For each point on the boundary there is a chain code A(i) that directs from the preset point to that point, and the chain code A(i1) directs to the point next to the current point. In this paper, an edge repair algorithm based on the combination of the eight-direction Freeman chain code and the difference of triple chain code is proposed. This method has a good repair effect on the loss of lung parenchyma edges, especially the vascular depression, and it is not easily affected by noise. The Freeman chain code is divided into eight-direction connection chain code and four-direction connection chain code, as shown in Fig. 8.

Fig. 8.

Fig. 8

Freeman chain code (a)four-direction connection chain code; (b)eight-direction connection chain .

The corresponding relationship between the chain code value and the boundary coordinates (x,y) when the eight-connection chain code describes the boundary is shown in Table 1.

Table 1.

Correspondence between the chain code value of eight neighborhood and the boundary coordinates (x, y).

x offset y offset The value of chain code
1 0 0
1 1 1
0 1 2
−1 1 3
−1 0 4
−1 −1 5
0 −1 6
1 −1 7

Rotating clockwise from the preset point to traverse the edges in the image. The value of the chain code increases by 1 for every 45°rotation, and the difference Diff(i) of the chain code is calculated by each point on the boundary. The threshold is set according to the different characteristics of chain code. When Diff(i)<1, the point is convex. When Diff(i)>2, the point is concave, and then the concave boundary of the lung parenchyma is repaired according to the convex and concave points. Ayerbe et al. [51] proposed Bresenham straight line algorithm to repair the defects of lung parenchyma. The repair effect is better for defects with small fissures in the lung parenchyma, while the outline of the lung parenchyma is generally curvilinear, which is not ideal for repairing large-area edge lesions. In this paper, a connection method based on the quadratic Bezier curve fitting interpolation is proposed. The primary Bezier curve is a line segment connecting the beginning and the end points, such as the line segment P0P2 in Fig. 9, and its expression is:

B(t)=(1t)P0+tP2,t[0,1] (33)

Fig. 9.

Fig. 9

Bezier curve interpolation defect .

The quadratic Bezier curve is a curve segment drawn by three points, and its expression is:

B(t)=(1t)2P0+2t(1t)P1+t2P2,t[0,1] (34)

Among them, an interpolation curve is required between point P0 and point P2. P1 is the directional control point at a certain distance from the midpoint of the line segment on the vertical bisector of the line P0P2. The distance is the optimal fixed value selected through testing, which can minimize the fitting interpolation error as much as possible. Through these three points, the curve interpolated to the position of the lung parenchyma will be drawn to complete the repair of lung parenchyma. Bezier curve patching is shown in Fig. 9.

The lung parenchyma template is repaired by Freeman chain code combined with quadratic Bezier curve can make up for the under-segmentation problem when repairing large-area defects with straight lines. At the same time, it has flexible adjustment to better meet the characteristics of lung parenchyma edge, which can achieve a good repair effect. Taking early COVID-19 as an example, the edge repair template of the lung parenchyma is shown in Fig. 11(e). Finally, the template is multiplied with the lung CT image, which can preserve a good effect of the preset ideal segmentation. It ensures the integrity of the lung parenchyma and do a good work for the subsequent detection of lung disease.

5. Analysis of experimental results

Aiming at the different kinds of COVID-19 CT images including early COVID-19, early asymptomatic COVID-19, resolution COVID-19, adenovirus pneumonia (suspected cases of COVID-19), staphylococcal pneumonia (suspected cases of COVID-19) are tested by the Intel E8200 CPU 2.5 GHz, RAM 8G, Matlab 2016a. The size of the image is 512 × 512, and the tested images and the golden standard segmented images of the lung parenchyma by doctors are shown in Fig. 10. The COVID-19 CT images tested in this paper are from COVID-19 CT segmentation dataset (https://medicalsegmentation.com/covid19/) and LIDC-IDRI (https://paperswithcode.com/dataset/lidc-idri). Two-dimensional entropy exhaustive segmentation(TDEE) [20], PSO two-dimensional entropy multi-threshold segmentation(PTEM) [52], standard firefly two-dimensional entropy multi-threshold segmentation (SFTM) [53], two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm(TDRMFA, paper method) are used for simulation comparison tests.

Fig. 10.

Fig. 10

Original COVID-19 CT images: (a)early COVID-19; (b)early asymptomatic COVID-19; (c)resolution COVID-19; (d)adenovirus pneumonia; (e)staphylococcal pneumonia; Golden standard segmented images of the lung parenchyma by doctors: (a1)early COVID-19; (b1)early asymptomatic COVID-19; (c1)resolution COVID-19; (d1)adenovirus pneumonia; (e1)staphylococcal pneumonia.

The experimental results show that the fixed threshold and the histogram threshold are easy to skip some areas adhered to the lung wall, and the segmentation effect is not ideal. If the threshold is not selected properly, it is easy to ignore the marginal nodules, and the gray contrast is not obvious, so it is impossible to observe the lesions comprehensively. The optimal threshold method proposed in this paper can perform the initial segmentation of the lungs more accurately. When the extracted lung parenchyma is repaired for defects, the convex hull algorithm cannot achieve the expected ideal effect when repairing the lung parenchyma template of large-area lesions. The rolling ball method cannot accurately determine the size of the defect when repairing the edge of the lung parenchyma template, and it is difficult to accurately set the radius of the sphere. If the radius is not selected properly, it is easy to have the problem of under segmentation or over segmentation. This paper improves this problem by calculating the Dice similarity coefficient (DSC) and sensitivity of image segmentation effect to verify the effectiveness of the experimental methods. The Dice similarity coefficient can be obtained by measuring the overlap between the segmentation results of the paper method and the manual segmentation results of experts, namely:

DSC=2×|uv||u|+|v| (35)

Where, u represents the segmentation result of this article, and v represents the golden standard segmentation result by doctor.

The sensitivity is obtained by measuring the ratio of the segmentation results of the paper method and the expert manual segmentation results to correctly segment the lung parenchyma area, namely:

Sensitivity=|uv||u| (36)

Where, u represents the segmentation result of this article, and v represents the golden standard segmentation result by doctor. The improved segmentation method in this paper has better indicators for the DSC and the sensitivity. The segmented lung parenchyma image can fully and clearly show the lesions, which is helpful to improve the accuracy and sensitivity of nodules detection. The extraction of lung parenchyma for different kinds of COVID-19 CT images are shown in Fig. 11, Fig. 12, Fig. 13, Fig. 14, Fig. 15.

Fig. 12.

Fig. 12

CT image of early asymptomatic COVID-19 segmented by TDRMFA: (a)binarized segmentation; (b)removal of trachea and bronchus; (c)removal of bed board and elimination of holes; (d)separation of the left and right lungs; (e)edge repair; (f)lung parenchyma segmentation. CT image of early asymptomatic COVID-19 segmented by SFTM: (a1)binarized segmentation; (b1)removal of trachea and bronchus; (c1) removal of bed board and elimination of holes; (d1)separation of the left and right lungs; (e1)edge repair; (f1)lung parenchyma segmentation. CT image of early asymptomatic COVID-19 segmented by PTEM: (a2)binarized segmentation; (b2)removal of trachea and bronchus; (c2)removal of bed board and elimination of holes; (d2)separation of the left and right lungs; (e2)edge repair; (f1)lung parenchyma segmentation. CT image of early asymptomatic COVID-19 segmented by PDEE: (a3)binarized segmentation; (b3) removal of trachea and bronchus; (c3)removal of bed board and elimination of holes; (d3)separation of the left and right lungs; (e3)edge repair; (f3)lung parenchyma segmentation.

Fig. 13.

Fig. 13

CT image of resolution COVID-19 segmented by TDRMFA: (a)binarized segmentation; (b)removal of trachea and bronchus; (c)removal of bed board and elimination of holes; (d)separation of the left and right lungs; (e)edge repair; (f)lung parenchyma segmentation. CT image of resolution COVID-19 segmented by SFTM: (a1)binarized segmentation; (b1)removal of trachea and bronchus; (c1) removal of bed board and elimination of holes; (d1)separation of the left and right lungs; (e1)edge repair; (f1)lung parenchyma segmentation. CT image of resolution COVID-19 segmented by PTEM: (a2)binarized segmentation; (b2)removal of trachea and bronchus; (c2)removal of bed board and elimination of holes; (d2)separation of the left and right lungs; (e2)edge repair; (f1)lung parenchyma segmentation. CT image of resolution COVID-19 segmented by PDEE: (a3)binarized segmentation; (b3) removal of trachea and bronchus; (c3)removal of bed board and elimination of holes; (d3)separation of the left and right lungs; (e3)edge repair; (f3)lung parenchyma segmentation.

Fig. 14.

Fig. 14

CT image of adenovirus pneumonia segmented by TDRMFA: (a)binarized segmentation; (b)removal of trachea and bronchus; (c)removal of bed board and elimination of holes; (d)separation of the left and right lungs; (e)edge repair; (f)lung parenchyma segmentation. CT image of adenovirus pneumonia segmented by SFTM: (a1)binarized segmentation; (b1)removal of trachea and bronchus; (c1) removal of bed board and elimination of holes; (d1)separation of the left and right lungs; (e1)edge repair; (f1)lung parenchyma segmentation. CT image of adenovirus pneumonia segmented by PTEM: (a2)binarized segmentation; (b2)removal of trachea and bronchus; (c2)removal of bed board and elimination of holes; (d2)separation of the left and right lungs; (e2)edge repair; (f1)lung parenchyma segmentation. CT image of adenovirus pneumonia segmented by PDEE: (a3)binarized segmentation; (b3) removal of trachea and bronchus; (c3)removal of bed board and elimination of holes; (d3)separation of the left and right lungs; (e3)edge repair; (f3)lung parenchyma segmentation.

Fig. 15.

Fig. 15

CT image of staphylococcal pneumonia segmented by TDRMFA: (a)binarized segmentation; (b)removal of trachea and bronchus; (c)removal of bed board and elimination of holes; (d)separation of the left and right lungs; (e)edge repair; (f)lung parenchyma segmentation. CT image of staphylococcal pneumonia segmented by SFTM: (a1)binarized segmentation; (b1)removal of trachea and bronchus; (c1) removal of bed board and elimination of holes; (d1)separation of the left and right lungs; (e1)edge repair; (f1)lung parenchyma segmentation. CT image of staphylococcal pneumonia segmented by PTEM: (a2)binarized segmentation; (b2)removal of trachea and bronchus; (c2)removal of bed board and elimination of holes; (d2)separation of the left and right lungs; (e2)edge repair; (f1)lung parenchyma segmentation. CT image of staphylococcal pneumonia segmented by PDEE: (a3)binarized segmentation; (b3) removal of trachea and bronchus; (c3)removal of bed board and elimination of holes; (d3)separation of the left and right lungs; (e3)edge repair; (f3)lung parenchyma segmentation.

This paper compares with the comparative methods and uses the accuracy IOU index to objectively evaluate the experimental results:

IOU=N11+N10N11+N10+N01+N00 (37)

Where, N11 represents the correctly segmented lung parenchyma region. N10 represents the correctly segmented background region. N01 represents the incorrectly segmented lung parenchyma region and N00 represents the incorrectly segmented background region. The greater the accuracy IOU, the more accurate the image segmentation is and the closer to the standard segmentation. The calculation results are shown in Table 2, Table 3, Table 4, Table 5, Table 6.

Table 2.

Comparison of TDRMFA with different methods for CT image of early COVID-19.

Segmented method Dice Sensitivity IOU
PDEE 0.7732 0.7898 0.8113
PTEM 0.8214 0.8342 0.8643
SFTM 0.8543 0.8623 0.8932
TDRMFA 0.9221 0.9353 0.9632

Table 3.

Comparison of TDRMFA with different methods for CT image of early asymptomatic COVID-19.

Segmented method Dice Sensitivity IOU
PDEE 0.8037 0.8145 0.8544
PTEM 0.8465 0.8530 0.8865
SFTM 0.8748 0.8865 0.9134
TDRMFA 0.9346 0.9484 0.9783

Table 4.

Comparison of TDRMFA with different methods for CT image of resolution COVID-19.

Segmented method Dice Sensitivity IOU
PDEE 0.7992 0.8067 0.8488
PTEM 0.8591 0.8662 0.8974
SFTM 0.8856 0.8982 0.9274
TDRMFA 0.9284 0.9334 0.9675

Table 5.

Comparison of TDRMFA with different methods for CT image of adenovirus pneumonia.

Segmented method Dice Sensitivity IOU
PDEE 0.7994 0.8038 0.8453
PTEM 0.8365 0.8430 0.8865
SFTM 0.8784 0.8885 0.9164
TDRMFA 0.9374 0.9474 0.9765

Table 6.

Comparison of TDRMFA with different methods for CT image of staphylococcal pneumonia.

Segmented method Dice Sensitivity IOU
PDEE 0.7738 0.7819 0.8259
PTEM 0.8183 0.8274 0.8679
SFTM 0.8547 0.8683 0.8982
TDRMFA 0.9193 0.9243 0.9584

Aiming at the different kinds of COVID-19 CT images, diversity-enhanced hybrid firefly algorithm (DeHFA) [54], uniform orthogonal firefly algorithm (UOFA) [55], chaos optimization firefly algorithm (COFA) [44], firefly algorithm based on the random mechanism and exponentially decreasing inertia weight (FARMEDI, paper method) are used for simulation comparison tests of segmentation parameters optimization. Moreover, aiming at the early COVID-19 CT image, we use the meta-heuristic algorithms to optimize image segmentation, including grasshopper optimization algorithm (GOA) [56], equilibrium optimization algorithm (EOA) [57], [58], marine predators algorithm (MPA) [59], [60] and traditional firefly algorithm (TFA). The fitness evolution curve for different COVID-19 CT images are shown in Fig. 16. The evaluation index of the optimization performance under different methods are shown in Table 7.

Fig. 16.

Fig. 16

Fitness evolution curve of segmented parameters optimization for different CT images: (a)Early COVID-19;(b)Early asymptomatic COVID-19;(c)Resolution COVID-19;(d)Adenovirus pneumonia;(e)staphylococcal pneumonia;(f)Early COVID-19.

Table 7.

Comparison of algorithm optimization performance.

CT image Algorithm Average Std Time processing
Early COVID-19 DeHFA 0.2853 0.0203 3.33
UOFA 0.4135 0.0582 3.47
COFA 0.5309 0.0947 4.71
FARMEDI 0.0901 0.0056 2.74

Early asymptomatic COVID-19 DeHFA 0.3898 0.0106 3.56
UOFA 0.5435 0.0508 3.82
COFA 0.6140 0.0614 4.46
FARMEDI 0.0640 0.0043 2.69

Resolution COVID-19 DeHFA 0.3731 0.0421 3.34
UOFA 0.6945 0.0458 3.76
COFA 0.8018 0.0751 4.48
FARMEDI 0.0517 0.0089 2.76

Adenovirus pneumonia DeHFA 0.2455 0.0116 3.46
UOFA 0.2508 0.0614 3.54
COFA 0.4335 0.0853 4.78
FARMEDI 0.0656 0.0034 2.82

Staphylococcal pneumonia DeHFA 0.3889 0.0391 3.35
UOFA 0.5345 0.0407 3.43
COFA 0.7018 0.0818 4.65
FARMEDI 0.0642 0.0047 2.89

Early COVID-19 MPA 0.2738 0.0018 3.27
EOA 0.3949 0.0052 3.39
GOA 0.5543 0.0086 4.82
TFA 0.8632 0.0083 5.29
FARMEDI 0.0901 0.0056 2.74

It can be seen from Fig. 16 and Table 7 that with the increase of the number of iterations, the standard deviation of the objective function value is gradually reduced and the processing time is gradually increased under different parameters optimization methods. Compared with the different comparison methods, the value of average and standard deviation for “FARMEDI” is still the smallest, and the processing time of “FARMEDI” is still the fastest. In conclusion, the paper method (FARMEDI) significantly improves the segmentation effect for different kinds of COVID-19 CT images of parameters optimization.

We compared the TDRMFA with 6 state-of-the-art approaches including Inf-Net [61], Chain code-SVM [62], CNN-Clustering [63], MLT [64], GAN-Unet [41], U-Net++ [65] to do segmentation simulations on the CT image of early COVID-19, and further comparing and analyzing the Dice, Sensitivity and IOU under different segmentation methods. The comparative CT image segmentation simulations of early COVID-19 are shown in Fig. 17. The evaluation index values are shown in Table 8.

Fig. 17.

Fig. 17

CT image of early COVID-19 (a)original CT image; (b)golden standard segmented image; (c)segmented by TDRMFA; (d)segmented by GAN-Unet; (e)segmented by U-Net++; (g)segmented by MLT; (h)segmented by CNN-Clustering; (i)segmented by Chain code-SVM; (j)segmented by Inf-Net.

Table 8.

Comparison of TDRMFA with segmented SOTA methods for CT image of early COVID-19.

Segmented method Dice Sensitivity IOU
Inf-Net 0.8494 0.8549 0.9254
Chain code-SVM 0.8519 0.8673 0.9372
CNN-Clustering 0.8626 0.8713 0.9443
MLT 0.8632 0.8784 0.9464
U-Net++ 0.8753 0.8814 0.9532
GAN-Unet 0.8786 0.8843 0.9536
TDRMFA 0.8872 0.9871 0.9667

From the change trend of the data in Table 8, compared with the comparative segmentation methods, TDRMFA has increased the value of Dice by about 0.0086, the value of Sensitivity has been increased by about 0.1028, the value of IOU has been increased by about 0.0131. It can be seen that the segmentation effect for the CT image of early COVID-19 under TDRMFA is the best. The early COVID-19 is characterized by the small number of lesion and the density is low. The TDRMFA method improves the segmentation accuracy for the CT image of early COVID-19 and reduces the missed diagnosis for the early lesions.

6. Conclusion

In view of the large computational and time-consuming problem for lung parenchymal segmentation of COVID-19, two-dimensional reciprocal cross entropy multi-threshold combined with improved firefly algorithm is proposed. The reciprocal cross entropy proposed in this paper solves the problem of the undefined and zero value of the logarithmic function. A weighted fuzzy threshold segmentation based on the spatial neighborhood information comprehensively considers the gray value and the spatial neighborhood information of image pixels, which improves the effect of image segmentation. In the process of calculating the optimal multi-threshold, the improved firefly algorithm is used to further improve the computational efficiency. The repair of lung parenchyma defects by Freeman chain code and Bezier curve is ideal, which can improve the accuracy of lung parenchyma segmentation and solve the problem of missed lesions. The experimental results for different kinds of COVID-19 CT images show that compared with the comparison methods, the performance of the paper method is the best. It not only segments COVID-19 lung parenchyma more accurately but also requires less running time. Thus, the time for doctors to sketch COVID-19 lesions is reduced, and it realizes the purpose of fully automatic segmentation of the COVID-19 lung parenchyma. However, there is a certain degree of difference in the structural characteristics of each person’s chest, so each lung parenchyma segmentation method has a limitation of robustness. In the segmentation results of the paper method, there is also a difference in segmentation accuracy for different CT images, which is the direction that needs to be researched in the future.

CRediT authorship contribution statement

Guowei Wang: Writing - original draft, Methodology, Software, investigation, validation. Shuli Guo: Conceptualization, Data curation. Lina Han: Writing - review & editing. Anil Baris Cekderi: Visualization.

Declaration of Competing Interest

No author associated with this paper has disclosed any potential or pertinent conflicts which may be perceived to have impending conflict with this work. For full disclosure statements refer to https://doi.org/10.1016/j.bspc.2022.103933.

Acknowledgments

This research has been partially funded by Beijing Municipal Science and Technology Commission-Beijing Natural Science Foundation, China(M21018) and project supported by National Key R&D Program of China (2017YFF0207400) and Beijing Natural Science Foundation-haidian District, China Joint Fund for original innovation(L192064).

References

  • 1.Wu Zunyou, McGoogan Jennifer M. Characteristics of and important lessons from the coronavirus disease 2019 (COVID-19) outbreak in China: Summary of a report of 72314 cases from the Chinese center for disease control and prevention. JAMA. 2020 doi: 10.1001/jama.2020.2648. [DOI] [PubMed] [Google Scholar]
  • 2.Mehta Puja, McAuley Daniel Francis, Brown Michael, Sanchez Emilie, Tattersall Rachel, Manson Jessica. COVID-19: consider cytokine storm syndromes and immunosuppression. Lancet (London, England) 2020;395:1033–1034. doi: 10.1016/S0140-6736(20)30628-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Fang Yicheng, Zhang Huangqi, Xie Jicheng, Lin Minjie, Ying Lingjun, Pang Peipei, Ji W-B. Sensitivity of chest CT for COVID-19: Comparison to RT-PCR. Radiology. 2020 doi: 10.1148/radiol.2020200432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bernheim Adam, Mei Xueyan, Huang Mingqian, Yang Yang, Fayad Zahi A., Zhang Ning, Diao Kaiyue, Lin Bin, Zhu Xiqi, Li Kunwei, Li Shaolin, Shan Hong, Jacobi Adam, Chung Michael S. Chest CT findings in coronavirus disease-19 (COVID-19): Relationship to duration of infection. Radiology. 2020 doi: 10.1148/radiol.2020200463. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Ardakani Ali Abbasian, Acharya U. Rajendra, Habibollahi Sina, Mohammadi Afshin. Covidiag: a clinical CAD system to diagnose COVID-19 pneumonia based on CT findings. Euro. Radiol. 2020:1–10. doi: 10.1007/s00330-020-07087-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Attallah Omneya, Ragab Dina Ahmed, Sharkas Maha A. MULTI-DEEP: A novel CAD system for coronavirus (COVID-19) diagnosis from CT images using multiple convolution neural networks. PeerJ. 2020;8 doi: 10.7717/peerj.10086. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Carfì Angelo, Bernabei Roberto, Landi Francesco. Persistent symptoms in patients after acute COVID-19. JAMA. 2020 doi: 10.1001/jama.2020.12603. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Soleimani Zahra, Soleimani Azam. Adrs due to COVID-19 in midterm pregnancy: successful management with plasma transfusion and corticosteroids. J. Maternal-Fetal Neonatal Medi. Official J. Euro. Assoc. Perinatal Med. 2020:1–4. doi: 10.1080/14767058.2020.1797669. the Federation of Asia and Oceania Perinatal Societies, the International Society of Perinatal Obstetricians. [DOI] [PubMed] [Google Scholar]
  • 9.Xiao Ai Tang, Tong Yi Xin, Zhang Sheng. False-negative of RT-PCR and prolonged nucleic acid conversion in COVID-19: Rather than recurrence. J. Med. Virol. 2020 doi: 10.1002/jmv.25855. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Arévalo-Rodriguez Ingrid, Buitrago-Garcia Diana, Simancas-Racines Daniel, Zambrano-Achig Paula, del Campo Rosa, Ciapponi Agustín, Sued Omar, Martínez-García Laura, Rutjes Anne Wilhelmina Saskia, Low Nicola, Pérez-Molina José Antonio, Zamora Javier. 2020. False-negative results of initial RT-PCR assays for COVID-19: a systematic review. MedRxiv. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Mardani Rajab, Vasmehjani Abbas Ahmadi, Zali Fatemeh, Gholami Alireza, Nasab Seyed Dawood Mousavi, Kaghazian Hooman, Kaviani Mehdi, Ahmadi Nayeb Ali. Laboratory parameters in detection of COVID-19 patients with positive RT-PCR; a diagnostic accuracy study. Arch. Acad. Emerg. Med. 2020;8 [PMC free article] [PubMed] [Google Scholar]
  • 12.Wang Yuhui, Dong Chengjun, Hu Yue, Li Chungao, Ren Qianqian, Zhang Xinling, Shi Heshui, Zhou Min. Temporal changes of CT findings in 90 patients with COVID-19 pneumonia: A longitudinal study. Radiology. 2020 doi: 10.1148/radiol.2020200843. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Li Kunhua, Wu Jiong, Wu Faqi, Guo Dajing, Chen Linli, Fang Zheng, Li Chuanming. The clinical and chest CT features associated with severe and critical COVID-19 pneumonia. Invest. Radiol. 2020 doi: 10.1097/RLI.0000000000000672. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Gozes Ophir, Frid-Adar Maayan, Greenspan Hayit, Browning Patrick D., Zhang Huangqi, Ji W.-B., Bernheim Adam, Siegel Eliot. 2020. Rapid AI development cycle for the coronavirus (COVID-19) pandemic: Initial results for automated detection & patient monitoring using deep learning CT image analysis. arXiv, arXiv:2003.05037. [Google Scholar]
  • 15.Vaishya Raju, Javaid Mohd., Khan Ibrahim Haleem, Haleem Abid. Artificial intelligence (AI) applications for COVID-19 pandemic. Diabetes Metab. Syndr. 2020;14:337–339. doi: 10.1016/j.dsx.2020.04.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Jin Shuo, Wang Bo, Xu Haibo, Luo Chuan, Wei Lai, Zhao Wei, Hou Xuexue, Ma Wenshuo, Xu Zhengqing, Zheng Zhuozhao, Sun Wenbo, Lan Lan, Zhang Wei, Mu Xiangdong, Shi Chenxi, Wang Zhongxiao, Lee Jihae, Jin Zijian, Lin Minggui, Jin Hong, Zhang Liang, Guo Jun, Zhao Benqi, Ren Zhizhong, Wang Shuhao, You Zheng, Dong Jiahong, Wang Xinghuan, Wang Jianming, Xu Wei. 2020. AI-assisted CT imaging analysis for COVID-19 screening: Building and deploying a medical AI system in four weeks. medRxiv. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Bai Harrison X., Wang Robin, Xiong Zeng, Hsieh Ben, Chang Ken, Halsey Kasey, Tran Thi My Linh, Choi Ji Whae, Wang Dong-Cui, Shi Lin-Bo, Mei Ji, Jiang Xiao-Long, Pan Ian, Zeng Qiuhua, Hu Ping-Feng, Li Yi-Hui, Fu Feixian, Huang Raymond Y, Sebro Ronnie A., Yu Qi-Zhi, Atalay Michael K., Liao Wei-Hua. AI augmentation of radiologist performance in distinguishing COVID-19 from Pneumonia of other etiology on chest CT. Radiology. 2020 doi: 10.1148/radiol.2020201491. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wang Shuihua, Zhu Ziquan, Zhang Yu-Dong. PSCNN: PatchShuffle convolutional neural network for COVID-19 explainable diagnosis. Front. Publ. Health. 2021;9 doi: 10.3389/fpubh.2021.768278. [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  • 19.Zhang Yu-Dong, Khan Muhammad Attique, Zhu Ziquan, Wang Shuihua. Pseudo zernike moment and deep stacked sparse autoencoder for COVID-19 diagnosis. Comput. Mater. Contin. 2021 [Google Scholar]
  • 20.Zhao Chen, Xu Yan, He Zhuo, Tang Jinshan, Zhang Yijun, Han Jungang, Shi Yuxin, Zhou Weihua. Lung segmentation and automatic detection of COVID-19 using radiomic features from chest CT images. Pattern Recognit. 2021;119:108071. doi: 10.1016/j.patcog.2021.108071. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Le Nguyen Quoc Khanh, Kha Quang Hien, Nguyen Van Hiep, Chen Yung-Chieh, Cheng Sho-Jen, Chen Cheng-Yu. Machine learning-based radiomics signatures for EGFR and KRAS mutations prediction in non-small-cell lung cancer. Int. J. Mol. Sci. 2021;22 doi: 10.3390/ijms22179254. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Le Nguyen Quoc Khanh, Hung Truong Nguyen Khanh, Do Duyen Thi, Lam Luu Ho Thanh, Dang Luong Huu, Huynh Tuan-Tu. Radiomics-based machine learning model for efficiently classifying transcriptome subtypes in glioblastoma patients from MRI. Comput. Biol. Med. 2021;132 doi: 10.1016/j.compbiomed.2021.104320. [DOI] [PubMed] [Google Scholar]
  • 23.Bhandari Ashish Kumar, Ghosh Arunangshu, Kumar Immadisetty Vinod. A local contrast fusion based 3D otsu algorithm for multilevel image segmentation. IEEE/CAA J. Autom. Sin. 2020;7:200–213. [Google Scholar]
  • 24.Shuanhu Wu, Hong Yan, DNA microarray image processing based on minimum error segmentation and histogram analysis, in: IS&T/SPIE Electronic Imaging, 2005.
  • 25.Upadhyay Pankaj, Chhabra Jitender Kumar. Kapur’s entropy based optimal multilevel image segmentation using crow search algorithm. Appl. Soft Comput. 2020;97 [Google Scholar]
  • 26.Raja Nadaradjane Sri Madhava, Sukanya S. Arockia, Nikita Y. Improved PSO based multi-level thresholding for cancer infected breast thermal images using otsu. Procedia Comput. Sci. 2015;48:524–529. [Google Scholar]
  • 27.Rajinikanth Venkatesan, Dey Nilanjan, Raj Alex Noel Joseph, Hassanien Aboul Ella, Santosh K.C., Raja Nadaradjane Sri Madhava. 2020. Harmony-search and otsu based system for coronavirus disease (COVID-19) detection using lung CT scan images. arXiv, arXiv:2004.03431. [Google Scholar]
  • 28.Xueqian Wang, Gang Li, Antonio J. Plaza, Adaptive Superpixel Segmentation with Fisher Vectors for Ship Detection in SAR Images, in: IGARSS 2020 - 2020 IEEE International Geoscience and Remote Sensing Symposium, 2020, pp. 1436–1439.
  • 29.Zhou Chengquan, Liang Dong, Yang Xiaodong, Xu Bo, Yang Guijun. Recognition of wheat spike from field based phenotype platform using multi-sensor fusion and improved maximum entropy segmentation algorithms. Remote. Sens. 2018;10:246. [Google Scholar]
  • 30.Estañón Carlos R., Aquino Norberto, Puertas-Centeno David, Dehesa Jesús S. Two-dimensional confined hydrogen: An entropy and complexity approach. Int. J. Quantum Chem. 2020 [Google Scholar]
  • 31.Chakraborty Rupak, Sushil Rama, Garg Madan Lal. An improved PSO-based multilevel image segmentation technique using minimum cross-entropy thresholding. Arab. J. Sci. Eng. 2018;44:3005–3020. [Google Scholar]
  • 32.Horng Ming-Huwi. Multilevel minimum cross entropy threshold selection based on the honey bee mating optimization. Expert Syst. Appl. 2010;37:4580–4592. [Google Scholar]
  • 33.Özyurt Fatih, Sert Eser, Avci Engin, Dogantekin Esin. Brain tumor detection based on convolutional neural network with neutrosophic expert maximum fuzzy sure entropy. Measurement. 2019;147 [Google Scholar]
  • 34.Hrosik Romana Capor, Tuba Eva, Dolicanin Edin, Jovanovic Raka, Tuba Milan. Brain image segmentation based on firefly algorithm combined with K-means clustering. Stud. Inf. Control. 2019 [Google Scholar]
  • 35.Wu Jinran, Wang You-Gan, Burrage Kevin, Tian Yu-Chu, Lawson Brodie A.J., Ding Zhe. An improved firefly algorithm for global continuous optimization problems. Expert Syst. Appl. 2020;149 [Google Scholar]
  • 36.Horng Ming-Huwi, Liou Ren-Jean. Multilevel minimum cross entropy threshold selection based on the firefly algorithm. Expert Syst. Appl. 2011;38:14805–14811. [Google Scholar]
  • 37.Alqazzaz Salma, Sun Xianfang, Yang Xin, Nokes Leonard Derek Martin. Automated brain tumor segmentation on multi-modal MR image using SegNet. Comput. Vis. Media. 2019;5:209–219. [Google Scholar]
  • 38.Ghaffari Reyhane, Golpardaz Maryam, Helfroush Mohammad Sadegh, Danyali Habibollah. A fast, weighted CRF algorithm based on a two-step superpixel generation for SAR image segmentation. Int. J. Remote Sens. 2020;41:3535–3557. [Google Scholar]
  • 39.Li Chenglong, Xia Weihao, Yan Yan, Luo Bin, Tang Jin. Segmenting objects in day and night: Edge-conditioned CNN for thermal image semantic segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2021;32:3069–3082. doi: 10.1109/TNNLS.2020.3009373. [DOI] [PubMed] [Google Scholar]
  • 40.Basukala Dibash, Jha Debesh, Kwon Goo-Rak. Brain image segmentation based on dual-tree complex wavelet transform and fuzzy C-means clustering algorithm. J. Med. Imag. Health Inform. 2018 [Google Scholar]
  • 41.Yan Wenjun, Wang Yuanyuan, Gu Shengjia, Huang Lu, Yan Fuhua, Xia Liming, Tao Qian. 2019. The domain shift problem of medical image segmentation and vendor-adaptation by unet-GAN. arXiv arXiv:1910.13681. [Google Scholar]
  • 42.He Hong, Tan Yonghong, Ying Jun, Zhang Wuxiong. Strengthen EEG-based emotion recognition using firefly integrated optimization algorithm. Appl. Soft Comput. 2020;94 [Google Scholar]
  • 43.Le Duc Thang, Bui Dac-Khuong, Ngo Tuan D., Nguyen Quoc-Hung, Nguyen-Xuan H. A novel hybrid method combining electromagnetism-like mechanism and firefly algorithms for constrained design optimization of discrete truss structures. Comput. Struct. 2019 [Google Scholar]
  • 44.Kaveh Ali, Javadi S.M. Chaos-based firefly algorithms for optimization of cyclically large-size braced steel domes with multiple frequency constraints. Comput. Struct. 2019 [Google Scholar]
  • 45.Ballesteros Javier, González-Estrada Octavio Andrés, Acevedo Heller Guillermo Sánchez. 2017. Damage detection in a unidimensional truss using the firefly optimization algorithm and finite elements. arXiv, arXiv:1706.04449. [Google Scholar]
  • 46.Ghosh parth, Mali Kalyani, Das Sitansu Kumar. Chaotic firefly algorithm-based fuzzy C-means algorithm for segmentation of brain tissues in magnetic resonance images. J. Vis. Commun. Image Represent. 2018;54:63–79. [Google Scholar]
  • 47.Dhal K.G., Das Arunita, Ray Swarnajit, Gálvez Jorge. Randomly attracted rough firefly algorithm for histogram based fuzzy image clustering. Knowl. Based Syst. 2021;216 [Google Scholar]
  • 48.Chakraborty Shouvik, Mali Kalyani. A morphology-based radiological image segmentation approach for efficient screening of COVID-19. Biomed. Signal Process. Control. 2021;69:102800. doi: 10.1016/j.bspc.2021.102800. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Ashour Amira S., Eissa Merihan M., Wahba Maram A., Elsawy Radwa A., Elgnainy Hamada Fathy, Tolba Mohamed Saeed, Mohamed Waleed S. Ensemble-based bag of features for automated classification of normal and COVID-19 CXR images. Biomed. Signal Process. Control. 2021;68:102656. doi: 10.1016/j.bspc.2021.102656. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Jia Dong Yao, Zhang Chuanwang, Wu Nengkai, Guo Zhigang, Ge Hairui. Multi-layer segmentation framework for cell nuclei using improved GVF Snake model, Watershed, and ellipse fitting. Biomed. Signal Process. Control. 2021;67 [Google Scholar]
  • 51.Ayerbe Alejandro Cosin, Patow Gustavo A. Clustered voxel real-time global illumination. Comput. Graph. 2022 [Google Scholar]
  • 52.Berta L., Rizzetto Francesco, Mattia Cristina De, Lizio Domenico, Felisi Marco, Colombo Paola Enrica, Carrazza Stefano, Gelmini Stefania, Bianchi L, Artioli D., Travaglini Francesca, Vanzulli Alberto, Torresin Alberto. Automatic lung segmentation in COVID-19 patients: Impact on quantitative computed tomography analysis. Physica Medica PM. 2021;87:115–122. doi: 10.1016/j.ejmp.2021.06.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Müller Dominik, Soto-Rey Iñaki, Kramer Frank. Robust chest CT image segmentation of COVID-19 lung infection based on limited data. Inform. Med. Unlocked. 2021;25:100681. doi: 10.1016/j.imu.2021.100681. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Aydilek Ibrahim Berkan, Karaçizmeli Izzettin Hakan, Tenekeci Mehmet Emin, Kaya Serkan, Gümüşçü Abdülkadir. Using chaos enhanced hybrid firefly particle swarm optimization algorithm for solving continuous optimization problems. Sadhana-Acad. Proc. Eng. Sci. 2021;46:1–22. [Google Scholar]
  • 55.Xu Sendren Sheng-Dong, Huang Hsu-Chih, Kung Yu-Chieh, Lin Shao-Kang. Collision-free fuzzy formation control of swarm robotic cyber-physical systems using a robust orthogonal firefly algorithm. IEEE Access. 2019;7:9205–9214. [Google Scholar]
  • 56.Dinh Phu-Hung. A novel approach based on grasshopper optimization algorithm for medical image fusion. Expert Syst. Appl. 2021;171 [Google Scholar]
  • 57.Dinh Phu-Hung. Multi-modal medical image fusion based on equilibrium optimizer algorithm and local energy functions. Appl. Intell. 2021;51:8416–8431. [Google Scholar]
  • 58.Dinh Phu-Hung. Combining gabor energy with equilibrium optimizer algorithm for multi-modality medical image fusion. Biomed. Signal Process. Control. 2021;68 [Google Scholar]
  • 59.Dinh Phu-Hung. A novel approach based on three-scale image decomposition and marine predators algorithm for multi-modal medical image fusion. Biomed. Signal Process. Control. 2021;67 [Google Scholar]
  • 60.Dinh Phu-Hung. An improved medical image synthesis approach based on marine predators algorithm and maximum gabor energy. Neural Comput. Appl. 2021:1–19. [Google Scholar]
  • 61.Fan Deng-Ping, Zhou Tao, Ji Ge-Peng, Zhou Yi, Chen Geng, Fu H., Shen Jianbing, Shao Ling. Inf-net: Automatic COVID-19 lung infection segmentation from CT images. IEEE Trans. Med. Imaging. 2020;39:2626–2637. doi: 10.1109/TMI.2020.2996645. [DOI] [PubMed] [Google Scholar]
  • 62.Kumar S. Pramod, Latte Mrityunjaya V. Lung parenchyma segmentation: Fully automated and accurate approach for thoracic CT scan images. IETE J. Res. 2018;66:370–383. [Google Scholar]
  • 63.Xu Mingjie, Qi Shouliang, Yue Yong, Teng Yueyang, Xu Lisheng, Yao Yudong, Qian Wei. Segmentation of lung parenchyma in CT images using CNN trained with the clustering algorithm generated dataset. BioMed. Eng. OnLine. 2019;18 doi: 10.1186/s12938-018-0619-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Liu Weihua, Liu Xiabi, Li Huiyu, Li Mincan, Zhao Xinming, Zhu Zheng. Integrating lung parenchyma segmentation and nodule detection with deep multi-task learning. IEEE J. Biomed. Health Inf. 2021;25:3073–3081. doi: 10.1109/JBHI.2021.3053023. [DOI] [PubMed] [Google Scholar]
  • 65.Ibtehaz N., Rahman Mohammad Sohel. MultiResUNet : Rethinking the U-net architecture for multimodal biomedical image segmentation. Neural Netw. 2020;121:74–87. doi: 10.1016/j.neunet.2019.08.025. [DOI] [PubMed] [Google Scholar]

Articles from Biomedical Signal Processing and Control are provided here courtesy of Elsevier

RESOURCES