Skip to main content
PLOS One logoLink to PLOS One
. 2025 Aug 13;20(8):e0329447. doi: 10.1371/journal.pone.0329447

Accuracy and real time optimization of remote sensing image change detection based on IRAU and DSC

Yingying Liu 1,*
Editor: Sohail Saif2
PMCID: PMC12349244  PMID: 40802607

Abstract

Image change detection is one of the important application branches of remote sensing technology in many fields. However, in complex environments, remote sensing image change detection is often subject to various interferences, resulting in low accuracy and poor real-time performance of detection results. The research focuses on the advantages and problems of residual networks and depth-wise separable convolution modules, designs a new remote sensing image change detection model, and finally sets up experiments for verification. The average accuracy of the proposed detection model before and after training convergence was 0.54 and 0.97. The accuracy of repeated detection ranged from 95.82% to 99.68%, and the area under curve of the model was 0.90. However, after removing the integrated residual attention unit and depth-wise separable convolution, the accuracy decreased by 1.91% and the latency increased by 117ms. In addition, the detection efficiency of the model for different elements ranged from 0.91 to 0.94, with high accuracy in detecting changes in spatial and temporal scales and small offsets. The actual accuracy and mean latency time of the model were 92.43% and 260ms, respectively. In summary, the proposed change detection model significantly improves the accuracy and real-time performance of remote sensing image processing, contributing to the expanded application of remote sensing dynamic detection technology in fields such as ocean monitoring and ecological research.

1. Introduction

Driven by technological development, the population has rapidly expanded, and the acquisition and processing of information at the geographic spatial scale have received great attention [1]. Remote sensing technology has wide coverage, fast data acquisition speed, large amount of information, and is not limited by ground conditions. It is widely used in environmental investigation and assessment, urban planning and management, which is also a research focus in computer vision [23]. Remote Sensing Image Change (RSIC) detection is a crucial application direction of remote sensing technology. According to the differences in detection ideas, there are pixel change detection and object change detection [45]. However, due to the increasing complexity of the Earth’s environment, traditional RSIC detection approaches have been unable to meet practical needs. For example, change detection methods based on object-oriented image analysis heavily rely on the segmentation results of image targets [6]. The automatic classification accuracy of time series change detection methods is extremely low and only used for large-scale target change analysis [7]. The three-dimensional change detection method requires high accuracy of the three-dimensional data itself and is computationally complex [8]. These issues seriously affect the accuracy and real-time performance of RSIC detection. Therefore, the research focuses on optimizing the effective acquisition and accurate detection of remote sensing images, focusing on the basic logic and excellent performance of modules such as Residual Networks (ResNets) and Depth-wise Separable Convolution (DSC), and making improvements. An Integrated Residual Attention Unit (IRAU) is proposed. Finally, a RSIC detection model on the basis of IRAU and DSC is constructed. The research aims to optimize the accuracy and real-time performance of RSIC detection models, enhance their resistance to environmental interference, and make positive contributions to the wider application, laying the foundation for promoting the rational planning and development of geographic spatial scales. The innovation lies in the organic combination of ResNet, attention mechanism and other modules. The IRAU unit is proposed and combined with DSC for improvement, establishing a new detection network architecture that effectively improves the detection accuracy and reduces detection latency.

The research is structured from four parts. Part one introduces the current research on RSIC detection worldwide. The second part starts from modules such as ResNets, Split and Concat (SPC), and DSC to establish an accurate and real-time RSIC detection model. The third part provides numerical examples and practical application analysis of the proposed image detection algorithm model to verify its reliability. The final section provides a comprehensive summary and analysis of the article.

2. Related works

With population growth and the large-scale development of various industries, the application of remote sensing technology in natural resources, agricultural production and other fields is rapidly increasing [910]. RSIC detection is the core of achieving geographic scale monitoring, which is also a crucial application direction that remote sensing technology needs to continuously expand and deepen [11]. However, in practical work, the performance of RSIC detection in complex environments is not stable, so many researchers are improving this problem [12]. C. Zhang et al. built a self-attention network on the basis of Siamese U-shaped structure to address the convolution operations being unable to capture global information, effectively improving the efficiency of processing detail information in RSIC [13]. G. Cheng et al. built an optimized separability deep learning network to address the boundary blurring between different semantic hierarchical features caused by traditional backbone networks. Through boundary maximization and directed attention mechanisms, the accuracy of RSIC detection was improved [14]. In response to the misalignment problem in traditional backbone networks, Y. Feng et al. designed an intra scale cross interaction and inter scale feature fusion network, which enhanced the integration performance of remote sensing image information at various resolutions by jointly extracting local global features [15]. In response to the low efficiency of Convolutional Neural Networks (CNN) in recognizing complex changes, W. Wang et al. designed a RSIC approach on the basis of multi-scale self-attention mechanism and mixed attention module, which improved the detection quality [16].

In addition, to address the sample imbalance in semantic change detection, Q. Zhu et al. built a twin global learning approach, which improved the accuracy and generalization of novel semantic change detection [17]. X. Zhang et al. designed an asymmetric cross attention hierarchical network that combined CNN and self-attention mechanisms to address issues such as poor interaction between CNN and self-attention mechanisms, resulting in significant computational resource consumption [18]. In response to the low accuracy and complex calculation of hyper-spectral remote sensing image detection, H. Firat et al. built a mixed 3D residual space spectral convolutional network to improve the accuracy of hyper-spectral remote sensing image classification [19]. In response to the limitations of convolutional layers on the global capture ability of neural networks, L. Wang et al. built a self-attention mechanism similar to a U-shaped network for urban scene segmentation, which improved the speed and accuracy of RSIC detection [20]. H. Chen et al. built an unsupervised multi-modal change detection framework on the basis of Fourier domain structural relationship analysis to address the difficulties in processing multi-modal remote sensing images caused by modal heterogeneity, which improved the fusion efficiency of local and non-local structural difference maps [21].

In summary, numerous researchers worldwide have noticed the problems in the operation of RSIC detection and have conducted multiple research efforts to address these issues. In addition, accurate and real-time RSIC detection is a prerequisite for expanding the use of remote sensing images at the geographic spatial scale, and its importance is self-evident. However, due to the redundancy of feature extraction information and unreasonable allocation of feature weights in traditional RSIC detection, the detection performance is unstable, and the accuracy and real-time performance of RSIC detection in complex environments are low. Therefore, based on the residual network, combined with attention mechanism and DSC module, a RSIC detection model on the basis of IRAU and DSC is established to improve the feature extraction accuracy and allocate the weights of detection results reasonably. The research aims to provide a comprehensive and innovative solution for addressing the latency and efficiency of RSIC detection in actual natural or artificial environments.

3. Methods and materials

With the expansion of industrial scale, the RSIC detection is also facing new challenges. A new RSIC detection model on the basis of IRAU and DSC is developed to address poor accuracy and real-time performance of traditional change detection techniques, which cannot meet diversified detection tasks. The model involves two parts, where IRAU optimizes the detection accuracy, while DSC optimizes the real-time performance of the model, jointly achieving efficient real-time detection of the model.

3.1. IRAU unit construction based on accuracy optimization

Accuracy and real-time performance are prerequisites for the application of RSIC detection in many aspects. However, traditional RSIC detection models are often affected by multiple factors, resulting in poor detection accuracy and real-time performance. A RSIC detection model on the basis of IRAU and DSC is designed to address the above issues. Among them, IRAU is responsible for optimizing model accuracy, which includes ResNet variants of -Res2Net + , SPC, and channel attention. By extracting semantic information from feature maps at various scales, IRAU enriches and refines the feature information to be detected, ultimately optimizing the accuracy of change detection. The structure of IRAU and Res2Net+modules is shown in Fig 1.

Fig 1. The structure of IRAU and Res2Net+ block [22].

Fig 1

In Fig 1(a), after the feature image is input into IRAU, it is processed through Res2Net+module, SPC module, and channel attention module to extract fine-grained semantic information of different scales. Finally, it is normalized, multiplied, and overlaid by softmax function to complete the image change detection output. In Fig 1 (b), the Finput is input into the Res2Net+module. First, the input is divided into four subsets after the first 1x1 convolution, defined as Xi and i=1,2,3,4, where each feature has the same scale size. In order to reduce parameters while increasing subsets, except for X1, all other sub-features have corresponding 3x3 convolution kernels, defined as Ki1, whose output is Yi, as displayed in equation (1).

{Yi=Xii=1Yi=Ki(Xi)i=2Yi=Ki(Xi+Yi1)2<i4 (1)

In equation (1), X1 is Y1. Y2 is the output of X2 after K2 convolution. When 2<i4 is present, all sub-features Xi are added to Ki1 and then input to Ki. After convolution, Yi is output. The channel shuffling operation is added to the Res2Net+module, and Yi is concatenated in the channel dimension. Then, the 1*1 convolution is performed to get FCS. Finally, FCS and Finput are added through residual connections to obtain Fres, which is beneficial for feature language stacking at different scales [2324]. In addition, the remaining sub-module structures of IRAU are shown in Fig 2.

Fig 2. The structure of SPC block and channel attention block [2526].

Fig 2

As shown in Fig 2 (a), the SPC module takes Fres as input and divides Fres into four parts, defined as Mi. i=0,1,2,3. SPC uses group convolution to process input tensors at different scales. The relationship between the convolution kernel size S and the size of the group convolution integration group g is g=2S12, and S=3,5,7,9. The SPC module obtains a multi-scale feature map Fi through group convolution processing, concatenates Fi to output Fspc, and independently learns multi-scale spatial information through each segmentation part [27]. In Fig 2 (b), the channel attention module takes Fspc as input and generates Favgc and Fmaxc feature maps through MaxPool and AvgPool operations. Then, Favgc and Fmaxc are forwarded to a Multi-Layer Perceptron (MLP) shared network to obtain a channel attention map Mc. The shared network is taken to each post feature map. The output feature vectors are fused through an element wise summation method, as shown in equation (2).

Mc(F)=σ(MLP(AvgPool(F))+MLP(maxPool(F)))=σ(W1(W0(Favgc))+W1(W0(Fmaxc))) (2)

Among them, σ signifies the sigmoid activation function. W0 and W1 represent the weights generated by MLP, which are shared for both Favgc and Fmaxc. The sigmoid activation function acts on W0. Based on the weight difference of feature vectors, the channel attention module assigns values to the input image features, which is beneficial for further processing in the future [28]. Finally, the weights W0 and W1 are normalized using IRAU’s softmax. The processed feature vector Mc(F) and feature map Fspc were pixel multiplied to obtain the output FIRA of IRAU. In summary, the second output FCS of Res2Net+module, the output Fspc of SPC module, and the final output FIRA of IRAU are shown in equation (3).

{FCS=ψ(Ω([Y1,Y2,Y3,Y4]))FSPC=SPC(Finput+FCS)FIRA=FSPC(CAM(FSPC)) (3)

In equation (3), ψ represents a 1*1 convolution operation. [] represents concatenation in the channel dimension. CAM signifies channel attention mechanism processing for features. represents the multiplication operation of feature vector pixels. To further optimize the accuracy of extracting multi-scale features and deep semantic information from RSIC detection models, a Multi-Scale Residual Block (MSRB) is used to extract image features [29]. Its structure is shown in Fig 3.

Fig 3. The structure of multi-scale residual block [30].

Fig 3

From Fig 3, the study replaces the more complex 5*5 and 7*7 convolution operations by concatenating three 3*3 convolutions to decline computational complexity. After the remote sensing image is input into MSRB, it undergoes three concatenated 3*3 convolution operations before being overlaid. The overlaid output features are then pixel added to the initial output after the 1*1 convolution operation. To capture spatial information in low-level features, a 3*3 convolutional kernel is used. To decline memory consumption, the quantity of filters in the internal convolutional layer is adjusted based on specific parameters, and a 1*1 convolutional kernel is used to ensure the accuracy of feature processing with a 3*3 convolutional kernel. To address the insignificant changes in model features and excessive irrelevant information in traditional detection, an Attention Gate (AG) module is introduced, as shown in Fig 4.

Fig 4. The structure of attention gate block [31].

Fig 4

From Fig 4, g signifies the input from the previous module, and x signifies the input from the skip connection. The skip connection is to skip one or several neural layers and transmit information to deeper layers of the neural network [32]. x and g are transformed into the same number of channels through a 1*1 convolution operation. After up-sampling, pixel addition is performed and rectified using a Rectified Linear Unit (ReLU) function. Then, another 1*1 convolution is conducted and the Sigmoid activation function is taken to get the importance score, which is assigned to each part of the feature map. The attention feature map is multiplied with the skip connection input to generate the final output of AG [33]. In summary, to ensure that the proposed IRAU can accurately and efficiently process the features of changing regions, a corresponding image network change detection twin network is set up, as shown in Fig 5.

Fig 5. Detection network structure with IRAU [34].

Fig 5

Fig 5 shows the process of using detection twin networks for detecting changes in images at different temporal phase. This network is based on the Siamese-Difference framework and ensures consistency in feature extraction of dual temporal images through shared weights. Each branch consists of MSRB and 2*2 max pooling layers, where MSRB is applied to get rich feature information. To optimize the detection accuracy of differences between images of different time phases, the network computes the absolute difference between the MSRB structure streams output by two branches and connects them into a difference feature map. Within the network, the AG module is introduced to combine the down-sampling feature map at the input and the up-sampling feature map at the output, aiming to highlight key changes in the image while suppressing noise. This mechanism emphases the real changing areas, improving the detection performance. The output design includes a combination of 2*2 deconvolution layers and MSRB, with the former responsible for restoring the resolution of the feature map and the latter further refining the features. Next, the feature map is passed to IRAU, which enhances the accuracy of edge recognition in changing regions through detailed multi-scale feature capture and adaptive channel optimization. It is crucial for generating high-quality change maps. Finally, after adjusting the quantity of channels in the feature map through a 1*1 convolutional layer, the change detection result is output.

3.2. Construction of dsc module and remote sensing image change detection model based on real-time optimization

The proposed IRAU and its application network have improved the accuracy of the detection model, but there are still challenges to the real-time performance. Due to factors such as climate and environment, traditional RSIC detection models have poor real-time performance and are difficult to complete continuous change detection tasks. Therefore, DSC is introduced into the proposed RSIC detection model, which is responsible for optimizing the real-time change detection performance. To improve the operational efficiency of DSC, the Efficient Channel Attention (ECA) is combined with Enhanced Atrous Spatial Pyramid Pooling (ASPP+) modules to construct an application network suitable for DSC operation. The DSC module and its application network structure are shown in Fig 6.

Fig 6. DSC block and its application network structure.

Fig 6

As shown in Fig 6 (a), DSC divides the convolutional block into two stages: deep convolution and point by point convolution. 3*3 deep convolution is responsible for performing a single filter convolution on each channel of the input image. Then, the batch normalization layer regularization [35] and ReLU activation operations are performed. Afterwards, the output channels of the depth convolution are combined by performing a 1*1 operation point by point convolution. The batch normalization layer regularization and ReLU activation are performed on the obtained feature maps again. In Fig 6 (b), the input of the DSC uses the Siamese-Difference framework to process images of different time phases in the detection area. DSC decline network parameters and computational complexity [36]. Therefore, the study uses DSC instead of conventional convolutional blocks for feature extraction at the input and output ends of its application network. In addition, the study adds ECA before skip connections to effectively integrate the rich spatial information extracted from shallow features at the input end with the rich semantic information extracted from deep features at the output end. Furthermore, to further optimize the detection performance of the overall detection model for targets of different scales in the changing area, an ASPP+module is introduced at the output end. After the feature map undergoes ASPP+operation, a 1*1 convolution is conducted again. Finally, the change map is output. The ECA module structure is shown in Fig 7.

Fig 7. The structure of efficient channel attention block [37].

Fig 7

ECA is an improvement based on compression and excitation networks, which uses local cross channel interaction strategies to decline the computational complexity while maintaining detection performance [38]. As shown in Fig 7, Global Average Pooling (GAP) is performed immediately after the feature map is input. Then, the fast one-dimensional convolution is used to perform cross channel fusion on kernel values, where k=5 represents the cross channel range. Afterwards, the channel weight values are obtained through Sigmoid activation operation, enabling the detection model to efficiently learn while avoiding the impact of dimensionality reduction on channel weight updates. ASPP is an algorithm model widely used in remote sensing image segmentation, which captures multi-scale information through dilated convolutions with various dilation rates to avoid detail loss caused by down-sampling. The proposed ASPP+ in the study is an improvement based on ASPP [39], and its structure is shown in Fig 8.

Fig 8. The structure of ASPP+ block.

Fig 8

As shown in Fig 8, in ASPP + , the dilated convolution with an original Rate value of 24 is replaced by a 1*1 convolution block. This is because in the traditional ASPP structure, as the Rate parameter of the dilated convolution increases, the effective area actually involved in the convolution operation in the input image will become smaller and smaller, which may lead to feature information loss and affect the detection performance [40]. After replacing with 1*1 convolutional blocks, the module can maintain a large receptive field while avoiding problems such as the effective convolution area being too small. The input feature map is subjected to DSC feature extraction to reduce the computational complexity. Subsequently, a 1*1 convolution is performed in parallel with three other dilated convolutions with different Rate values of 6, 12, and 18, respectively, to capture feature information of various scales in the feature map. Then, the feature maps generated by the above operations are concatenated to obtain a multi-scale feature fusion feature map. Finally, in order to further optimize the feature information, the concatenated feature map undergoes another 1*1 convolution operation, which helps reduce redundant information and achieve efficient fusion of multi-scale features, thereby enhancing the detection capability of the model.

In summary, a RSIC detection network model on the basis of IRAU and DSC is proposed to address the accuracy and real-time issues of change detection models. This model is divided into two parts: IRAU and DSC, aimed at optimizing the detection accuracy and efficiency. The workflow is as follows. Remote sensing images from various time periods are input into two branches of the network, and multi-scale feature extraction and feature map dimensionality reduction are achieved through shared weight MSRB and 2*2 max pooling layers. The feature map is handled by the DSC module to reduce parameters and computational costs. The AG module integrates input down-sampling and output up-sampling feature maps to enhance change features and remove redundant information. The output end first restores the resolution of the feature map through a 2*2 deconvolution layer, and then jumps and connects with the input end feature map to fuse multi-level features. Then, the feature map is further refined by MSRB and sent to the IRAU module for multi-scale feature extraction and adaptive feature refinement. Finally, the feature map is adjusted for the number of channels through a 1*1 convolutional layer to produce a change map. This model effectively improves the accuracy of change detection while declining computational complexity, making it suitable for applications in high-resolution remote sensing images.

4. Results

To verify the effectiveness and superiority of the RSIC detection model on the basis of IRAU and DSC, the theoretical basis and algorithm analysis mentioned above are comprehensively studied. The research chooses a deep learning framework on the basis of PyTorch as the experimental platform to conduct model training and simulation testing experiments on different algorithms. In addition, the study also conducts experiments on the change detection performance in practical environments, collects relevant data, analyzes the experimental results in detail to compare the model performance.

4.1. Model training and simulation testing experiment

In model training and simulation testing experiments, Windows 10 is chosen as the operating system and AdamW optimizer is used for model training and optimization. The study selects Feature Pyramids within the Network, Deformable Component Detection Model, You Only Look Once, and Deep Image Fusion Network as comparative methods, named M1, M2, M3, and M4, respectively. The research uses the change detection dataset as a training and test set (8:2) with remotely sensed data from a region in Angers, France (Geographical distribution center: Latitude: 47° 37’ 28“ N, Longitude: 000° 09’ 55” W) covering the period from June 2019 to September 2021, containing Landsat-8 and Sentinel-2 multi-temporal imagery with a resolution of 10–30 meters. The data sources are the U.S. Geological Survey and European Space Agency public databases, covering typical scenarios such as urban sprawl and vegetation changes. The study first trains the model to verify its training efficiency, as presented in Fig 9.

Fig 9. Algorithm training accuracy experiments.

Fig 9

According to Fig 9 (a), IRAU-DSC completed convergence at 1700 iterations, with the fastest convergence speed, followed by M1, which completed convergence at 2200 iterations. M3 and M4 converged at 2400 and 3000 iterations, respectively, while M2 had the slowest convergence speed and completed it at 3700 iterations. From Fig 9 (b), before and after convergence, M2 had the lowest average accuracy, at 0.25 and 0.68, respectively, followed by M4, with average accuracy before and after convergence of 0.36 and 0.83, respectively. The accuracy of M1 and M3 before and after convergence was between 0.36–0.75 and 0.87–0.96, respectively, while IRAU-DSC had the highest accuracy before and after convergence, with mean values of 0.54 and 0.97. The IRAU-DSC has better learning performance than other methods and the highest accuracy. Afterwards, to verify the stability of the trained algorithm in detecting images, repeated recognition experiments are conducted, as presented in Fig 10.

Fig 10. Repeat the detection experiment.

Fig 10

According to Fig 10 (a), the detection accuracy of M2 varied the most, ranging from 67.53% to 76.94%, followed by M4, with accuracy fluctuating between 83.12% and 88.93%. The detection accuracy of M1 and M3 showed relatively small changes, with maximum amplitudes of 2.68% and 5.81%, respectively, while the accuracy of IRAU-DSC showed the smallest changes, ranging from 95.82% to 99.68%. According to Fig 10 (a), in terms of average accuracy, IRAU-DSC had the highest accuracy, reaching 97.56%. Next was M1, with an average accuracy of 94.68%. The average accuracy of M3 and M4 was slightly lower, at 90.91% and 85.91% respectively, while M2 had the lowest accuracy, with an average of only 71.97%. After training on the dataset, the performance of IRAU-DSC is more stable, and the accuracy is consistent with the above training results. Subsequently, the study conducts Receiver Operating Characteristic curve (ROC) and Area Under Curve (AUC) analysis, as shown in Fig 11.

Fig 11. Receiver operating characteristic curves for different algorithms.

Fig 11

From Fig 11 (a) and (b), in the ROC, the false positive rate increased the fastest when it was between 0–0.3, and the true positive rate slowed down when it was between 0.3–0.6, and gradually stabilized after 0.6. When the false positive rate was the same, M2 had the lowest true positive rate, followed by M4. At this point, the true positive rates of M1 and M3 were slightly higher than M4, but IRAU-DSC had the highest true positive rate. In addition, there was a similar trend in the AUC values of different algorithm models. The AUC value of IRAU-DSC was the highest, reaching 0.90, followed by M1 with an AUC of 0.85. The AUC of M3 and M4 was 0.82 and 0.77, with M2 having the lowest AUC value of only 0.70. The IRAU-DSC has higher change detection performance and fully meets the needs of remote sensing detection tasks. And Kappa coefficient can reflect the degree of consistency between the classification results and the real labeling, while excluding the consistency caused by random factors. Therefore, the study compares the Kappa coefficients of different methods to verify their accuracy for change detection maps, and the results are shown in Table 1.

Table 1. Kappa coefficients for different algorithms.

Methods TP TN FP FN Kappa Coefficient
M1 700 200 50 50 0.75
M2 720 180 70 30 0.78
M3 680 220 30 70 0.72
M4 750 150 100 0 0.82
IRAU-DSC 800 180 20 0 0.9

In Table 1, the Kappa coefficient of IRAU-DSC reaches 0.90, which is significantly higher than that of M1 (0.75), M2 (0.78), M3 (0.72) and M4 (0.82), and its TP (800) and TN (180) are optimal and FP (20) and FN (0) are the lowest, which verifies that the model has a high accuracy and anti-jamming ability in the recognition of changing regions. Compared with traditional methods such as in-network feature pyramid and deformable part detection, the proposed algorithm effectively improves the robustness and discriminative consistency of change detection in complex scenes through deep feature fusion strategy. Experimental results show that the accuracy of IRAU-DSC for change detection graph is much higher than other methods. To further validate the importance of the proposed IRAU and DSC models in detection tasks, ablation experiments are conducted on the detection models, as displayed in Table 2.

Table 2. Ablation experiments of various block.

algorithms Block Accuracy
(%)
Time
Deviation(ms)
MSRB AG IRAU ECA ASPP+ DSC
M1 × × × × × 93.25 1775
M2 × × × × × × 70.34 2846
M3 × × × × × × 91.13 2169
M4 × × × × × × 86.08 2198
IRAU-DSC × × × × 92.19 1130
IRAU-DSC × × × 96.07 1129
IRAU-DSC × × × × 91.25 471
IRAU-DSC × × × 94.16 236
IRAU-DSC × × 96.51 299
IRAU-DSC 98.42 182

According to Table 2, the detection accuracy of M1, M2, M3, and M4 ranged from 70.34% to 93.25%, with the latency of 1775ms, 2846ms, 2169ms, and 2198ms, respectively. At this time, the detection accuracy of IRAU-DSC was 98.42%, with the latency of 182ms. When the detection model proposed in the study only had MSRB and AG, its accuracy was 92.19% and the latency was 1130ms. After introducing IRAU, its accuracy increased by 3.88% and the latency remained almost unchanged. When the detection model only had ECA and ASPP + , its accuracy was 91.25% and the latency was 471ms. After introducing DSC, the accuracy increased by 2.91% and the latency decreased by 235ms. In addition, when IRAU-DSC lost IRAU and DSC, its accuracy decreased by 1.91% and latency increased by 117ms. The introduced module exerts a crucial role in the overall detection model, and the proposed improvements to IRAU and DSC are the core for achieving excellent performance of the model.

4.2. Experimental study on the performance of model actual change detection

Based on the above results, the study compares M1 and M2 algorithms, with IRAU-DSC as the research object. The detection performance in the time environment is analyzed to verify its potential for practical application and promotional value. The study first selects houses, rivers, and roads, and compares their actual performance through different algorithms to detect different elements. The results are shown in Fig 12.

Fig 12. Confusion matrix analysis for different algorithms.

Fig 12

As shown in Fig 12 (a), M1 had a higher overall detection efficiency than M2, but lower than IRAU-DSC. M1 had the highest detection efficiency for rivers, at 0.88, and for houses and roads, at 0.83 and 0.86 respectively. As shown in Fig 12 (b), M2 had the lowest detection efficiency in practical applications, ranging from 0.76 to 0.80, and M2 also had good detection efficiency for rivers. According to Fig 12 (c), IRAU-DSC had high detection performance for houses, rivers, and roads, with an overall efficiency of over 0.91. Among them, the algorithm had the highest detection efficiency for roads, reaching 0.94. The experimental results show that in practical applications, IRAU-DSC has better detection efficiency for different elements than other algorithms and can perform remote sensing detection in complex environments. Then, to verify the performance in detecting changes at the spatial scale, the study selects images of a large courtyard at different time phases for change detection experiments, as presented in Fig 13.

Fig 13. Change detection on a spatial scale.

Fig 13

(Source from https://earthexplorer.usgs.gov/; Latitude: 47° 37’ 28“ N, Longitude: 000° 09’ 55” W).

From Figs 13 (a) and (b), the main change in this courtyard was the dismantling of the tool room in the upper right corner, which was converted into a sun drying area. There were more temporary material storage areas in the middle, and larger warehouses were developed on the barren land in the lower part. In addition, there were significant changes in the number of cars in two different temporal images. From Fig 13 (c), IRAU-DSC detected changes in the tool room and drying area in the upper right corner, with a slight deviation in the detection of changes in the temporary stacking area in the middle. It performed well in detecting changes in the lower warehouse and detected 17 changes in vehicles. From Fig 13 (d), M1 did not detect the dismantling of the tool room, and the temporary material stacking area and warehouse in the middle and lower parts were found to be distorted. Nine changes in the vehicles were detected, but some of the detections showed deviations. From Fig 13 (e), M2 experienced severe distortion when detecting tool room, drying area, and warehouse, and did not detect any changes in the temporary stacking area. M2 also detected changes in 7 vehicles, but there were significant deviations. To further verify the real-time detection performance on a continuous time scale, the study performs continuous change detection on a white car driving at a certain intersection. The results are shown in Fig 14.

Fig 14. Change detection on a continuous time scale.

Fig 14

(Source from: https://pixabay.com/zh/videos/cars-roundabout-rondo-flowers-road-168811/).

From Fig 14 (a), at the 3rd second of the screen, the car entered the screen and traveled along the right side. At this time, IRAU-DSC and M1 detected the position of the target car, but M1 showed a significant deviation, while M2 erroneously detected the gray car on the right side. From Fig 14 (b), at the 11th second of the screen, when the car reached the upper position of the screen, IRAU-DSC still correctly detected the target position, while the deviation of M1 increased and the detection latency of M2 was larger. The detection result was the target vehicle position around the 10th second. From Fig 14 (c), at the 17th second of the screen, when the car reached the left edge of the screen, there was a slight deviation in IRAU-DSC detection. The detection deviation of M1 was further increased, and the detection result of M2 still had a latency, failing to correctly detect the position of the current target car. The IRAU-DSC has high real-time and accuracy in detecting continuous changes of moving targets. To verify the authenticity of the actual detection efficiency experiment mentioned above, the accuracy and latency time of algorithm detection in practical application scenarios are calculated, as presented in Table 3.

Table 3. The actual change detection efficiency of algorithms.

Number of experiments Accuracy(%) Time deviation(ms)
M1 M2 IRAU-DSC M1 M2 IRAU-DSC
1 84.47 68.73 93.34 1535 2574 211
2 87.95 70.78 92.00 1722 2768 207
3 85.34 68.25 93.94 1469 2499 302
4 86.96 65.46 91.93 1347 2559 281
5 85.79 80.49 90.27 1360 2966 318
6 83.67 70.74 90.56 1279 2435 253
7 84.77 66.11 93.68 1194 2706 296
8 92.65 67.05 92.72 1370 2722 202
9 85.74 70.58 92.38 1427 2578 257
10 85.58 70.18 93.43 1610 2743 273
Mean 86.29 69.84 92.43 1431 2655 260

According to Table 3, in practical application scenarios, the detection efficiency decreased. M2 had the lowest accuracy, ranging from 65.46% to 80.49%, and the highest latency, ranging from 2435ms to 2966ms. M1 had slightly higher efficiency, with accuracy and latency ranging from 83.67% to 92.65% and 1194ms-1722ms, respectively. The accuracy and latency of IRAU-DSC were still superior to other algorithms, with accuracy ranging from 90.27% to 93.94%, with an average of 92.43%, and latency ranging from 202ms to 318ms, with an average of 260ms. The experimental results show that in practical environmental applications, the detection efficiency calculation data of IRAU-DSC is consistent with the detection results, and the experimental results are true and reliable. The excellent performance of IRAU-DSC comes from the proposed IRAU units and the optimization improvement of IRAU and DSC. In addition, in order to verify the generalization and robustness of the research model in different spectral scenarios, the study selects different spectral scenarios for change detection experiments, and the results are shown in Table 4.

Table 4. Comparison of model performance in different different spectral scenarios.

Scene conditions Methods Accuracy (%) Time Deviation (ms) Kappa
Coefficient
Hyperspectral
(224 bands)
M1 86.2 34.5 0.81
M2 88.7 41.2 0.83
IRAU-DSC 93.5 28.7 0.9
Multi-spectral
(8 bands)
M1 82.4 28.1 0.76
M2 84.9 33.6 0.79
IRAU-DSC 89.3 24.9 0.86
Low resolution multispectral
(4 bands)
M1 78.6 21.3 0.71
M2 80.2 25.8 0.73
IRAU-DSC 85.4 19.5 0.81

In Table 4, the study validates the model performance in hyperspectral (224 bands), multispectral (8 bands), and low-resolution multispectral (4 bands) scenarios.IRAU-DSC achieves the optimal Accuracy (93.5%) and Kappa (0.90) in hyperspectral scenarios, with an improvement of 6.8%/4.8% and 0.09/0.07 over M1/M2, and a reduction of 16 ms to 12 ms in the temporal bias (28.7 ms). It maintains 85.4% accuracy and 0.81 Kappa for low-resolution data, and improves 5.2%−7.2% accuracy and 0.08–0.10 Kappa compared with the baseline model, and the computational efficiency is stable (19.5ms), which indicates that its adaptive spectral feature fusion mechanism can effectively deal with the band difference and noise interference, and can balance the accuracy and real-time performance. It shows that its adaptive spectral feature fusion mechanism can effectively cope with band differences and noise interference, and take into account both accuracy and real-time performance. Finally, in order to verify the consistency of the model detection results with the ground data, the study selects the Intersection over Union (IoU), the ratio of correctly categorized pixels (changed/unchanged) to the total pixels (Correct Pixel Ratio), and the proportion of undetected in the real change area (Omission Error), Position Error and Area Error are compared, and the results are shown in Table 5.

Table 5. Consistency verification of model detection results with ground data.

Metric IoU (%) Correct pixel ratio
(%)
Omission error
(%)
Position error (pixel) Area error
(%)
M1 72.3 86.1 18.5 4.7 12.8
M2 75.8 88.7 14.2 3.9 9.6
IRAU-DSC 82.6 93.4 8.3 2.1 5.3

In Table 5, the study verifies the consistency of the detection results with the ground data through IoU, CPR and other indicators. Among them, the IoU of IRAU-DSC reaches 82.6%, which is 10.3% and 6.8% higher than that of M1 (72.3%) and M2 (75.8%); and its Correct pixel ratio (93.4%) is significantly higher than that of the baseline model (86.1%/88.7%). Its Correct pixel ratio (93.4%) is significantly higher than that of the baseline model (86.1%/88.4%). Moreover, the omission error (8.3%) and location error (2.1 pixels) of IRAU-DSC are reduced by 6.2%−9.4% and 1.8 pixels-2.6 pixels, respectively, and the area error (5.3%) is reduced by 7.5%/4.3% compared with that of the M1/M2 model, which proves that its detection boundary, spatial location, and area coverage are highly matched with the real area of the change, which effectively guarantees the change detection results. It effectively guarantees the ground verifiability of the change detection results.

5. Discussion and conclusion

In response to the low accuracy and poor real-time performance of traditional remote sensing change detection models, this study proposed the IRAU unit and improved it by combining modules such as DSC. Finally, the IRAU-DSC model was proposed, which improved the detection performance by increasing the receptive field and reasonable weighting values. During training on the dataset, the average accuracy of IRAU-DSC before and after convergence was 0.54 and 0.97, respectively, while the accuracy ranges of other algorithms before and after convergence were 0.15–0.75 and 0.66–0.96, respectively. After training, the repeat detection accuracy of IRAU-DSC ranged from 95.82% to 99.68%, while the average accuracy of other algorithms were 94.68%, 71.97%, 90.91%, and 85.91%, respectively. In ROC analysis, the AUC value of IRAU-DSC was 0.90, while the AUC values of other algorithms were between 0.70 and 0.85. In the ablation experiment, after losing IRAU and DSC, the accuracy of the model decreased by 1.91% and the latency increased by 117ms. In addition, in actual detection, the detection efficiency of IRAU-DSC was above 0.91, while the efficiency of other algorithms was between 0.76–0.88. In spatial scale detection experiments, IRAU-DSC was more accurate in detecting changes in size within space, while other algorithms suffered from low detection accuracy, distortion, or deviation. In continuous detection experiments, IRAU-DSC showed high accuracy and real-time performance, while other algorithms encountered issues such as error detection, drift, and latency. In the actual detection efficiency calculation, the average latency time of IRAU-DSC was 260ms, while the latency time of other algorithms was between 1194ms-2966ms. In summary, the research has practical application value in improving the accuracy and real-time performance of RSIC detection. However, IRAU relies on multiple modules connected in series with jump connections, which may lead to gradient flow redundancy and reduced feature fusion efficiency. And although DSC reduces the computational volume, the sensory field is limited in high-resolution input, and the fine-grained feature extraction is insufficient. To address the above problems, the research will simplify the module interaction path to optimize the gradient propagation efficiency, design adaptive cavity convolution to enhance the receptive field of DSC, introduce a cross-module dynamic weight allocation mechanism to improve the compatibility of heterogeneous features, and explore the lightweight architecture to balance the computational load and detection accuracy.

Supporting information

S1 File. Minimal data set definition.

(DOC)

pone.0329447.s001.doc (92.5KB, doc)

Data Availability

All relevant data are within the paper and its Supporting Information files.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Pal S, Roy A, Palaiahnakote S, Pal U. Adapting a Swin Transformer for License Plate Number and Text Detection in Drone Images. AIA. 2023;1(3):145–54. doi: 10.47852/bonviewaia3202549 [DOI] [Google Scholar]
  • 2.Casagli N, Intrieri E, Tofani V, Gigli G, Raspini F. Landslide detection, monitoring and prediction with remote-sensing techniques. Nat Rev Earth Environ. 2023;4(1):51–64. doi: 10.1038/s43017-022-00373-x [DOI] [Google Scholar]
  • 3.Wang Y, Albrecht CM, Braham NAA, Mou L, Zhu XX. Self-Supervised Learning in Remote Sensing: A review. IEEE Geosci Remote Sens Mag. 2022;10(4):213–47. doi: 10.1109/mgrs.2022.3198244 [DOI] [Google Scholar]
  • 4.Bai T, Wang L, Yin D, Sun K, Chen Y, Li W, et al. Deep learning for change detection in remote sensing: a review. Geo-spatial Information Science. 2022;26(3):262–88. doi: 10.1080/10095020.2022.2085633 [DOI] [Google Scholar]
  • 5.Liu Z, Xu J, Liu M, Yin Z, Liu X, Yin L, et al. Remote sensing and geostatistics in urban water-resource monitoring: a review. Mar Freshw Res. 2023;74(10):747–65. doi: 10.1071/mf22167 [DOI] [Google Scholar]
  • 6.Suganyadevi S, Seethalakshmi V, Balasamy K. A review on deep learning in medical image analysis. Int J Multimed Inf Retr. 2022;11(1):19–38. doi: 10.1007/s13735-021-00218-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Li G, Jung JJ. Deep learning for anomaly detection in multivariate time series: Approaches, applications, and challenges. Information Fusion. 2023;91:93–102. doi: 10.1016/j.inffus.2022.10.008 [DOI] [Google Scholar]
  • 8.Stilla U, Xu Y. Change detection of urban objects using 3D point clouds: A review. ISPRS Journal of Photogrammetry and Remote Sensing. 2023;197:228–55. doi: 10.1016/j.isprsjprs.2023.01.010 [DOI] [Google Scholar]
  • 9.Cavender-Bares J, Schneider FD, Santos MJ, Armstrong A, Carnaval A, Dahlin KM, et al. Integrating remote sensing with ecology and evolution to advance biodiversity conservation. Nat Ecol Evol. 2022;6(5):506–19. doi: 10.1038/s41559-022-01702-5 [DOI] [PubMed] [Google Scholar]
  • 10.Cheng C, Zhang F, Shi J, Kung H-T. What is the relationship between land use and surface water quality? A review and prospects from remote sensing perspective. Environ Sci Pollut Res Int. 2022;29(38):56887–907. doi: 10.1007/s11356-022-21348-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Zhu Q, Guo X, Li Z, Li D. A review of multi-class change detection for satellite remote sensing imagery. Geo-spatial Information Science. 2022;27(1):1–15. doi: 10.1080/10095020.2022.2128902 [DOI] [Google Scholar]
  • 12.Wang D, Zhang J, Du B, Xia G-S, Tao D. An Empirical Study of Remote Sensing Pretraining. IEEE Trans Geosci Remote Sensing. 2023;61:1–20. doi: 10.1109/tgrs.2022.3176603 [DOI] [Google Scholar]
  • 13.Zhang C, Wang L, Cheng S, Li Y. SwinSUNet: Pure Transformer Network for Remote Sensing Image Change Detection. IEEE Trans Geosci Remote Sensing. 2022;60:1–13. doi: 10.1109/tgrs.2022.3160007 [DOI] [Google Scholar]
  • 14.Cheng G, Wang G, Han J. ISNet: Towards Improving Separability for Remote Sensing Image Change Detection. IEEE Trans Geosci Remote Sensing. 2022;60:1–11. doi: 10.1109/tgrs.2022.3174276 [DOI] [Google Scholar]
  • 15.Feng Y, Xu H, Jiang J, Liu H, Zheng J. ICIF-Net: Intra-Scale Cross-Interaction and Inter-Scale Feature Fusion Network for Bitemporal Remote Sensing Images Change Detection. IEEE Trans Geosci Remote Sensing. 2022;60:1–13. doi: 10.1109/tgrs.2022.3168331 [DOI] [Google Scholar]
  • 16.Wang W, Tan X, Zhang P, Wang X. A CBAM Based Multiscale Transformer Fusion Approach for Remote Sensing Image Change Detection. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2022;15:6817–25. doi: 10.1109/jstars.2022.3198517 [DOI] [Google Scholar]
  • 17.Zhu Q, Guo X, Deng W, Shi S, Guan Q, Zhong Y, et al. Land-Use/Land-Cover change detection based on a Siamese global learning framework for high spatial resolution remote sensing imagery. ISPRS Journal of Photogrammetry and Remote Sensing. 2022;184:63–78. doi: 10.1016/j.isprsjprs.2021.12.005 [DOI] [Google Scholar]
  • 18.Zhang X, Cheng S, Wang L, Li H. Asymmetric Cross-Attention Hierarchical Network Based on CNN and Transformer for Bitemporal Remote Sensing Images Change Detection. IEEE Trans Geosci Remote Sensing. 2023;61:1–15. doi: 10.1109/tgrs.2023.3245674 [DOI] [Google Scholar]
  • 19.Firat H, Asker ME, Bayindir Mİ, Hanbay D. 3D residual spatial–spectral convolution network for hyperspectral remote sensing image classification. Neural Comput & Applic. 2022;35(6):4479–97. doi: 10.1007/s00521-022-07933-8 [DOI] [Google Scholar]
  • 20.Wang L, Li R, Zhang C, Fang S, Duan C, Meng X, et al. UNetFormer: A UNet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery. ISPRS Journal of Photogrammetry and Remote Sensing. 2022;190:196–214. doi: 10.1016/j.isprsjprs.2022.06.008 [DOI] [Google Scholar]
  • 21.Chen H, Yokoya N, Chini M. Fourier domain structural relationship analysis for unsupervised multimodal change detection. ISPRS Journal of Photogrammetry and Remote Sensing. 2023;198:99–114. doi: 10.1016/j.isprsjprs.2023.03.004 [DOI] [Google Scholar]
  • 22.Chen F, Wang N, Yu B, Wang L. Res2-Unet, a New Deep Architecture for Building Detection From High Spatial Resolution Images. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2022;15:1494–501. doi: 10.1109/jstars.2022.3146430 [DOI] [Google Scholar]
  • 23.Gao X, Xu L, Wang F, Hu X. Multi-branch aware module with channel shuffle pixel-wise attention for lightweight image super-resolution. Multimedia Systems. 2022;29(1):289–303. doi: 10.1007/s00530-022-00976-w [DOI] [Google Scholar]
  • 24.Huang K, Xu Z. Lightweight video salient object detection via channel-shuffle enhanced multi-modal fusion network. Multimed Tools Appl. 2023;83(1):1025–39. doi: 10.1007/s11042-023-15251-x [DOI] [Google Scholar]
  • 25.Zhang M, Sun L, Li X, Zhu W, Yi Y, Yan F. A multi-scale self-calibrating lung nodule detection based on SPC-UNet. Biomedical Signal Processing and Control. 2025;104:107515. doi: 10.1016/j.bspc.2025.107515 [DOI] [Google Scholar]
  • 26.Wan D, Lu R, Shen S, Xu T, Lang X, Ren Z. Mixed local channel attention for object detection. Engineering Applications of Artificial Intelligence. 2023;123:106442. doi: 10.1016/j.engappai.2023.106442 [DOI] [Google Scholar]
  • 27.Zhai W, Gao M, Anisetti M, Li Q, Jeon S, Pan J. Group-split attention network for crowd counting. J Electron Imag. 2022;31(04). doi: 10.1117/1.jei.31.4.041214 [DOI] [Google Scholar]
  • 28.Lee HS, Cho SI. Locally Adaptive Channel Attention-Based Spatial–Spectral Neural Network for Image Deblurring. IEEE Trans Circuits Syst Video Technol. 2023;33(10):5375–90. doi: 10.1109/tcsvt.2023.3250509 [DOI] [Google Scholar]
  • 29.Fu H, Liang F, Liang J, Li B, Zhang G, Han J. Asymmetric Learned Image Compression With Multi-Scale Residual Block, Importance Scaling, and Post-Quantization Filtering. IEEE Trans Circuits Syst Video Technol. 2023;33(8):4309–21. doi: 10.1109/tcsvt.2023.3237274 [DOI] [Google Scholar]
  • 30.Liang H, Cao J, Zhao X. Multi-scale dynamic adaptive residual network for fault diagnosis. Measurement. 2022;188:110397. doi: 10.1016/j.measurement.2021.110397 [DOI] [Google Scholar]
  • 31.Yu M, Chen X, Zhang W, Liu Y. AGs-Unet: Building Extraction Model for High Resolution Remote Sensing Images Based on Attention Gates U Network. Sensors (Basel). 2022;22(8):2932. doi: 10.3390/s22082932 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Tran T-T, Pham V-T. Fully convolutional neural network with attention gate and fuzzy active contour model for skin lesion segmentation. Multimed Tools Appl. 2022;81(10):13979–99. doi: 10.1007/s11042-022-12413-1 [DOI] [Google Scholar]
  • 33.Tomar NK, Jha D, Riegler MA, Johansen HD, Johansen D, Rittscher J, et al. FANet: A Feedback Attention Network for Improved Biomedical Image Segmentation. IEEE Trans Neural Netw Learn Syst. 2023;34(11):9375–88. doi: 10.1109/TNNLS.2022.3159394 [DOI] [PubMed] [Google Scholar]
  • 34.He X, Chen Q, Tang L, Wang W, Liu T, Li L, et al. Federated Continuous Learning Based on Stacked Broad Learning System Assisted by Digital Twin Networks: An Incremental Learning Approach for Intrusion Detection in UAV Networks. IEEE Internet Things J. 2023;10(22):19825–38. doi: 10.1109/jiot.2023.3282648 [DOI] [Google Scholar]
  • 35.Liu A, Tang S, Chen X, Huang L, Qin H, Liu X, et al. Towards Defending Multiple $$\ell _p$$-Norm Bounded Adversarial Perturbations via Gated Batch Normalization. Int J Comput Vis. 2023;132(6):1881–98. doi: 10.1007/s11263-023-01884-w [DOI] [Google Scholar]
  • 36.Junayed MS, Islam MB, Imani H, Aydin T. PDS-Net: A novel point and depth-wise separable convolution for real-time object detection. Int J Multimed Info Retr. 2022;11(2):171–88. doi: 10.1007/s13735-022-00229-6 [DOI] [Google Scholar]
  • 37.Karthik R, Vaichole TS, Kulkarni SK, Yadav O, Khan F. Eff2Net: An efficient channel attention-based convolutional neural network for skin disease classification. Biomedical Signal Processing and Control. 2022;73:103406. doi: 10.1016/j.bspc.2021.103406 [DOI] [Google Scholar]
  • 38.Kang DH, Cha Y-J. Efficient attention-based deep encoder and decoder for automatic crack segmentation. Struct Health Monit. 2022;21(5):2190–205. doi: 10.1177/14759217211053776 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Cui Z, Lei Y, Wang Y, Yang W, Qi J. Hand gesture segmentation against complex background based on improved atrous spatial pyramid pooling. J Ambient Intell Human Comput. 2022;14(9):11795–807. doi: 10.1007/s12652-022-03736-w [DOI] [Google Scholar]
  • 40.Fu D, Zeng X, Han S, Lin H, Li W. Nested Densely Atrous Spatial Pyramid Pooling and Deep Dense Short Connection for Skeleton Detection. IEEE Trans Human-Mach Syst. 2023;53(1):75–84. doi: 10.1109/thms.2022.3224552 [DOI] [Google Scholar]

Decision Letter 0

Sohail Saif

21 Apr 2025

PONE-D-24-59365Accuracy and Real time Optimization of Remote Sensing Image Change Detection based on IRAU and DSCPLOS ONE

Dear Dr. Liu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Jun 05 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Sohail Saif, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1.Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that Figure 13 and 14 in your submission contain satellite images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth). For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

 We require you to either (a) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (b) remove the figures from your submission:

 a. You may seek permission from the original copyright holder of Figure 13 and 14 to publish the content specifically under the CC BY 4.0 license. 

 We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

 In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

 b. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

The following resources for replacing copyrighted map figures may be helpful:

 USGS National Map Viewer (public domain): http://viewer.nationalmap.gov/viewer/

The Gateway to Astronaut Photography of Earth (public domain): http://eol.jsc.nasa.gov/sseop/clickmap/

Maps at the CIA (public domain): https://www.cia.gov/library/publications/the-world-factbook/index.html and https://www.cia.gov/library/publications/cia-maps-publications/index.html

NASA Earth Observatory (public domain): http://earthobservatory.nasa.gov/

Landsat: http://landsat.visibleearth.nasa.gov/

USGS EROS (Earth Resources Observatory and Science (EROS) Center) (public domain): http://eros.usgs.gov/#

Natural Earth (public domain): http://www.naturalearthdata.com/

4. Please remove your figures from within your manuscript file, leaving only the individual TIFF/EPS image files, uploaded separately. These will be automatically included in the reviewers’ PDF.

5. We note that your Data Availability Statement is currently as follows: All relevant data are within the manuscript and in Supporting Information files.

Please confirm at this time whether or not your submission contains all raw data required to replicate the results of your study. Authors must share the “minimal data set” for their submission. PLOS defines the minimal data set to consist of the data required to replicate all study findings reported in the article, as well as related metadata and methods (https://journals.plos.org/plosone/s/data-availability#loc-minimal-data-set-definition).

For example, authors should submit the following data:

- The values behind the means, standard deviations and other measures reported;

- The values used to build graphs;

- The points extracted from images for analysis.

Authors do not need to submit their entire data set if only a portion of the data was used in the reported study.

If your submission does not contain these data, please either upload them as Supporting Information files or deposit them to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories.

If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. If data are owned by a third party, please indicate how others may request data access

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: No

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: 1. Please explain about the Data properly. Remote Sensing data must include about the location of the study area, time of data acquisition and also about data source.

2. Reference citation in methodology is very less. Except your results all the figure captions should be cited with proper References.

3. Your Figure 13. must be represented in the form of a map with proper legends highlighting the changes.

4. ROC curve will only give the model Accuracy, but what is the Accuracy of your result? (the Change Detection map)

5. You must calculate "Kappa Statistic" or "Kappa Coefficient" for your result accuracy.

6. What is the validation of your result with Ground Data?

Reviewer #2: Your research idea is promising, but I would suggest elaborating on certain aspects.

1.Since your focus is on remote sensing data, have you considered using different types of remote sensing data(Hyperspectral,multispectral) to support and justify your proposed approach.

2.Please include a description of the dataset you used in your study.

3.Briefly state the limitations of your proposed method.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: Yes:  Mr. Siddhartha Bhaumick

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Review.doc

pone.0329447.s002.doc (22KB, doc)

Decision Letter 1

Sohail Saif

17 Jul 2025

Accuracy and Real time Optimization of Remote Sensing Image Change Detection based on IRAU and DSC

PONE-D-24-59365R1

Dear Dr. Liu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Sohail Saif, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: All comments have been addressed. The manuscript supports the conclusion, and it is technically sound. Hence, it is ok and recommended for publication.

Reviewer #2: All the questions have been addressed successfully.The idea of this work is unique, In future you may apply all the methods with different Datasets.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: No

Reviewer #2: No

**********

Acceptance letter

Sohail Saif

PONE-D-24-59365R1

PLOS ONE

Dear Dr. Liu,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Sohail Saif

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Minimal data set definition.

    (DOC)

    pone.0329447.s001.doc (92.5KB, doc)
    Attachment

    Submitted filename: Review.doc

    pone.0329447.s002.doc (22KB, doc)
    Attachment

    Submitted filename: Response to reviewers comments.docx

    pone.0329447.s003.docx (65.6KB, docx)

    Data Availability Statement

    All relevant data are within the paper and its Supporting Information files.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES