Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Dec 9;16(12):e0258681. doi: 10.1371/journal.pone.0258681

Water tank and swimming pool detection based on remote sensing and deep learning: Relationship with socioeconomic level and applications in dengue control

Higor Souza Cunha 1,*, Brenda Santana Sclauser 1, Pedro Fonseca Wildemberg 2, Eduardo Augusto Militão Fernandes 2, Jefersson Alex dos Santos 2, Mariana de Oliveira Lage 3, Camila Lorenz 4, Gerson Laurindo Barbosa 5, José Alberto Quintanilha 6, Francisco Chiaravalloti-Neto 4
Editor: Guy J-P Schumann7
PMCID: PMC8659416  PMID: 34882711

Abstract

Studies have shown that areas with lower socioeconomic standings are often more vulnerable to dengue and similar deadly diseases that can be spread through mosquitoes. This study aims to detect water tanks installed on rooftops and swimming pools in digital images to identify and classify areas based on the socioeconomic index, in order to assist public health programs in the control of diseases linked to the Aedes aegypti mosquito. This study covers four regions of Campinas, São Paulo, characterized by different socioeconomic contexts. With mosaics of images obtained by a 12.1 MP Canon PowerShot S100 (5.2 mm focal length) carried by unmanned aerial vehicles, we developed deep learning algorithms in the scope of computer vision for the detection of water tanks and swimming pools. An object detection model, which was initially created for areas of Belo Horizonte, Minas Gerais, was enhanced using the transfer learning technique, and allowed us to detect objects in Campinas with fewer samples and more efficiency. With the detection of objects in digital images, the proportions of objects per square kilometer for each region studied were estimated by adopting a Chi-square distribution model. Thus, we found that regions with low socioeconomic status had more exposed water tanks, while regions with high socioeconomic levels had more exposed pools. Using deep learning approaches, we created a useful tool for Ae. aegypti control programs to utilize and direct disease prevention efforts. Therefore, we concluded that it is possible to detect objects directly related to the socioeconomic level of a given region from digital images, which encourages the practicality of this approach for studies aimed towards public health.

1. Introduction

1.1 Motivations to develop the study

Aedes aegypti is the main vector of dengue, Zika, and chikungunya, among other arboviruses [1, 2]. In the past, this mosquito species has been responsible for major urban yellow fever epidemics in Brazil. Though Ae. aegypti was considered eradicated from the country in the 1950s, the mosquito reemerged in the 1970s [3]. As a result, dengue epidemics have been occurring in Brazil since the 1980s with increasing intensity and an increasing number of serious cases and deaths [4]. Since 2015, epidemics of Zika (associated with cases of microcephaly) and chikungunya have also been affecting the entire Brazilian territory. In addition to being a vector of these diseases, the Ae. aegypti can transmit other arboviruses, such as the Mayaro virus [5], and has the potential to reintroduce yellow fever in urban areas [2].

Several studies have shown a relationship between higher levels of Ae. aegypti infestation and dengue risk in areas with lower socioeconomic levels [610]. Thus, identification of areas at greatest risk of the presence of the vector and the occurrence of the diseases is one of the measures adopted to optimize vector and disease control. Zambon et al. [11] and Carlucci et al. [12] used Google Earth images to detect swimming pools in Mediterranean European cities and found an association between areas with higher socioeconomic levels and higher densities of swimming pools. This density could be used as a proxy for class segregation [12] or to classify areas according to their socioeconomic levels. This is an example of a fast, up-to-date, and partially or fully automated classification of areas according to socioeconomic levels that could be useful to identify priority areas for the development of vector control. If a higher density of swimming pools could be a good indicator of areas with better socioeconomic levels, then a higher density of water tanks installed on roofs, which are very common in less privileged areas in Brazil [13], could be a good indicator for identifying areas with lower socioeconomic levels. Niebergall et al. [14] and Ayush et al. [15] used remote sensing images to detect water tanks and other targeted objects to investigate vulnerability and predict poverty in urban areas; thus, both swimming pools and water tanks could be used for the socioeconomic characterization of areas.

Another way to improve vector and disease control is the detection of key breeding sites for vector infestation; depending on their conditions water tanks and swimming pools could be characterized as key breeding sites. The main issue with water tanks is the inadequate use of tank screens, allowing vector access and breeding. This is more prominent in poorer areas. As for swimming pools, they could become significant breeding grounds for mosquitoes when they are not properly treated [16]. Several studies have already highlighted that the use of geographic information systems (GIS), spatial analysis tools, and remote sensing could be beneficial in improving arbovirus surveillance and control [1621]. Additionally, the use of remote sensing and artificial intelligence (AI) are promising approaches, especially for the purpose of identifying water tanks on roofs and swimming pools to classify areas based on their socioeconomic class or to identify vector breeding sites [22].

1.2 Remote sensing

Remote sensing is a quick and low-cost method to characterize and identify Ae. aegypti breeding sites and classify the areas according to different degrees of risk. Some studies have used this technique to identify risk areas for vector-borne diseases or have at least highlighted the possibility [16]. The studies developed by Sanchez et al. [18] and Lorenz et al. [21], using satellite images, identified a relationship between the number of Ae. aegypti adult females and the presence of asbestos slabs and roofs in the building.

Some studies have applied machine learning (ML) to identify soil cover using random forest (RF) [23] and support vector machine (SVM) [24] techniques. The recent development of AI techniques and ML methods based on deep learning (DL), which has revolutionized the recognition of patterns in images and established the state of the art in various applications including remote sensing, can be an important technique for optimizing of the use of aerial images in the surveillance and control of arboviruses.

1.3 Deep learning in computer vision

Computer vision is a field of study that enables computers to obtain information with a high level of abstraction from images and videos. A major motivation for this field is the automation of tasks that require analyzing substantial amounts of visual data in a short period of time [25]. In this context, one of the most useful tasks is object detection, where the computer not only detects objects present in an image, but also their location, standing out in a variety of applications inside and outside of computer science. Similarly, ML is a branch of computer science focused on making predictions through a mathematical model inferred by the machine itself, based on the training data it receives as an input. Such a technique is essential in applications in computer vision since it is often infeasible for a human to solve the problems present in this field.

A deep neural network (DNN) is composed of multiple layers of perceptrons, which are small learning units. The function of a perceptron is to learn weights for a linear decision function. The DNN can combine multiple perceptron layers to create an effective decision function that models the class of interest [25]. For pattern recognition in images, the most common DNN method is to use multiple layers of convolutional perceptrons. This allows the network to effectively learn color, shape, and texture patterns that help the model to differentiate the classes of interest. For more information about convolutional neural networks (CNN), [2527].

In the context of vector control and remote sensing, the application of DL can be highly effective, as can be seen in Zhang et al. [28], since it allows processing large geographic areas such as urban areas (cities) by allowing the pattern recognition associated with the development of Ae. aegypti breeding sites, for example. DL methods are often used in the application of wide-area information because it has high potential in the extraction of characteristics, such as texture, color, and format. Therefore, this study uses DL to detect objects most correlated to the infestation of the vector. In this context, the objects were chosen based on one of the most important elements in the life cycle of this species of mosquito, i.e., water.

1.4 Related work

Several studies have used satellite imagery to detect swimming pools in urban areas [16, 22, 30, 33, 34, 37]. More specifically, some of these studies made use of shallow learning methods to fulfill the task, while others utilized a variety of different computer vision techniques. Moreover, the motivations for such studies range from helping emergency services to controlling disease vectors.

In 2008, a group of researchers from Australia published a paper describing a technique to identify swimming pools in satellite images using SVM [29], intending to help emergency services locate water bodies throughout the city to facilitate bushfire fights in their country. In the approach presented, as well as in the work herein, only the Red, Green, and Blue (RGB) bands from the images in the dataset were used, calculating the difference between the blue-red and blue-green values, and feeding this data to the SVM. Unfortunately, no metric was calculated to determine the effectiveness of the method [30]. In 2009, another paper tried to identify swimming pools by using a different approach: first, the RGB color space of the image was converted to the C1, C2, and C3 [31], although only the C1 component was used, and a threshold was applied to generate a binary image containing only the detected pools. Additionally, all the regions considered too small for a typical swimming pool were eliminated. Thereafter, contours of the detected objects were refined through a snake algorithm [32]. Finally, the authors claimed that their method was able to detect more than 93% of the swimming pools contained in the dataset assembled by them [33].

Other studies used the normalized difference water index (NDWI) [34], which was proposed to delineate open water features in aerial imagery by making use of the near-infrared and green bands of the given image. A study in 2011 combined NDWI with the rectangular fit space metric [35] to detect swimming pools to help control the population of the Culex mosquito vector of the West Nile Virus in the United States of America [22], showing a user’s accuracy of 80.1% in pool positive samples and 92.8% in negative ones. A similar method using NDWI and a simple threshold was proposed to tackle the same problem, achieving better results with a user’s accuracy of 98% in parcels with pools and 88.3% in those without them [16]. In 2014, the normalized difference swimming pools index (NDSPI) was proposed and used to detect pools semi-autonomously, alongside principal component analysis (PCA) [36], image segmentation, and region adjacency graphs (RAGs), obtaining an overall accuracy of 99.86% [37].

Some studies also investigated methods for the detection of water tanks. Niebergall et al. [14] developed new image processing techniques and GIS integration methodologies to assess the vulnerability within megacities based on remote sensing information and showed that smaller water tanks, which are often located on the top of a building’s roof, could be identified well in the panchromatic band. Saini et al. [38] used faster region-based CNN (Faster R-CNN) for detecting overhead water tanks from satellite imagery. Ayush et al. [15] used an interpretable computational framework to predict poverty at the local level by applying object detectors at high resolution, and one of the multiple objects used was water tanks.

1.5 Objectives of the study

This study aimed to use remote sensing and DL to detect swimming pools and water tanks installed on roofs in areas of Campinas, São Paulo (SP) to investigate the relationship between the presence of these objects and the area’s socioeconomic level and to discuss the possible application of the obtained results in the control of dengue and its vector, Ae. aegypti. It should be noted that, although mosquitoes might breed even in insignificant amounts of water [2, 39], the present study is limited to water tanks and swimming pools as they are more easily detectable using high-resolution imagery. In the case of this study, the imagery comprises images obtained by a 12.1 megapixel (MP) Canon PowerShot S100 (5.2 mm focal length) carried by an unmanned aerial vehicle (UAV) with 0.03 m/pixel spatial resolution. This study has the potential to improve dengue control with remote sensing by detecting specific targets and generate an updated classification in real time for socioeconomic indexes. This classification is considerably important in countries like Brazil, where the demographic census, which is the main source of knowledge about the living conditions of the population, is only conducted every 10 years. Furthermore, this methodology would contribute to identifying areas at the greatest risk, which, once prioritized, could make disease control more effective. By using automated methods, the areas and places most likely to encounter this type of breeding site could be found more efficiently. This would allow public services to act directly on properties that could be considered key for maintenance and increase of vector infestation.

This paper is organized as follows. Section 2 presents the Materials and Methods, which introduces the study area, describes how the datasets were built, describes all the tools and software that we used, and outlines the methodology we used to train the neural network. The results we obtained are provided in Section 3, while in Section 4 our results are discussed by analyzing the socioeconomic relationship that we discovered, describing our limitations during the research, describing how this study can be expanded to other areas, and denoting our possible next steps and future research. Finally, Section 5 provides our conclusions.

2. Materials and methods

2.1 Study area and datasets

2.1.1 Belo Horizonte

In Fernandes, Wildemberg & Dos Santos [40], the authors assembled two separate datasets for the city of Belo Horizonte, MG, Southeast Brazil (Fig 1) at 19°48’57"S and longitude 43°57’15"W. The climate is subtropical and wet, with an average annual temperature and rainfall of 20.8°C and 1205 mm, respectively.

Fig 1. City of Belo Horizonte with the neighborhoods used in the dataset.

Fig 1

One of the datasets had annotated pools (BH-Pools), and the other had annotated water tanks (BH-WaterTanks). Both datasets consisted of imagery from several neighborhoods of the city of Belo Horizonte, Brazil, the data were acquired through the Google Earth Pro tool. The RGB images were exported from a flying altitude of 330 m above the ground with a size of 3840 × 2160 pixels (4K). Each 4K image was then cropped into six patches of 1280 × 1080 pixels each, and patches without any annotations were removed. Subsequently, 80% of the images from each neighborhood were used as a training dataset, and the other 20% were used as a test dataset.

The dataset BH-Pools consisted of 200 4K images from eight different neighborhoods (25 images for each) and contained 3980 annotated pools. The data preparation step resulted in 655 patches designated for training and 160 for testing. The dataset BH-WaterTanks was made up of 150 4K images from six neighborhoods (25 images for each) and contained 16216 annotated water tanks. The data preparation step resulted in 608 patches designated for training and 148 for testing. Because the images are from Google Earth Pro, it is not possible to determine their exact spatial resolution, since each one is a combination of several different images from different observations that are meshed together to form a new, higher-resolution image. However, through a visual analysis of the imagery, one can estimate their spatial resolution to be around 0.15 m/pixel.

2.1.2 Campinas

In a manner similar to that employed for the Belo Horizonte datasets, two new datasets were assembled with imagery from the city of Campinas, SP, Southeast Brazil (Fig 2), located at 22°53’03"S and 47°02’39"W. The city has the third-largest population in the state with just over one million inhabitants and has a high rate of human development (0.805). The climate is hot and temperate, with an average annual temperature and rainfall of 19.3°C and 1315 mm, respectively.

Fig 2. City of Campinas with neighborhoods used in the dataset.

Fig 2

Four areas were selected (Fig 2) with different socioeconomic levels [41], which were characterized using the information of the demographic census of 2010 made available by the Brazilian Institute of Geography and Statistics (IBGE). Area 1 had the best socioeconomic level, with an average income per household of 1,807 Reais (the Brazilian currency) and 3.0 inhabitants per household; Area 4 had the worst socioeconomic level, with an average income per household of 755 Reais and 3.5 inhabitants per household. Areas 2 and 3 had intermediate socioeconomic levels, with 1,285 and 1,138 Reais of average income per household, respectively, and 3.2 and 3.5 inhabitants per household, respectively.

We also considered the Social Vulnerability Index of SP (IPVS) to characterize these four areas; the index was based on the 2010 demographic census and made available for the census tracts. IPVS was created by the State System of Data Analysis Foundation (SEADE) [42, 43] and measures the population’s vulnerability to poverty. It has two dimensions, a socioeconomic and a demographic one, and considers variables such as the per capita household income, the percentage of women aged 10 to 29 in charge of the households, and the state of subnormal agglomeration (slums) in the census tract. Areas 1 (22°56’31"S; 47°05’57’’W) and 2 (22°58’01’’S; 47°08’41’’W) were classified as areas with mostly exceptionally low vulnerability (median and high socioeconomic levels and young, adult, and elderly families). Area 3 (22°58’14’’S; 47°07’58’’W) was classified with exceptionally low, low, and medium vulnerability (low socioeconomic levels and young families). Area 4 (22°59’29’’S; 47°07’15’’W) was classified with mostly extremely high vulnerability (low socioeconomic levels and young families living in slums). Areas 1 to 4 had 377, 689, 624, and 543 households and occupied areas of 0.32, 0.30, 0.41, and 0.27 km2, respectively.

The study on the four areas described above required the approval of the Ethics Committee in Plataforma Brasil (Brazil Platform System), according to the presentation certificate for ethical evaluation no. 43813015.9.0000.0059 and approved according to opinion no. 1.082.780.

We cropped the UAV images of the four areas into smaller patches (240 × 240-pixel clippings). With these patches, two datasets were assembled, one for swimming pools and another for water tanks. The swimming pool dataset was assembled by labeling 74 patches, which were divided into 52 samples for training and validation, and 22 for testing. We labeled 146 patches to assemble the water tanks dataset, which was divided into 102 for training and validation, and 44 for testing.

Table 1 presents a comparison between the Belo Horizonte and Campinas datasets. It can be seen that Dataset 2 presents a small number of samples (patches) for a DL application, because we only had mosaic images of four regions in Campinas and there was not an abundance of swimming pools and water tanks, unlike in Belo Horizonte.

Table 1. Comparison between the Belo Horizonte and Campinas datasets.
Dataset 1 Dataset 2
City Belo Horizonte, Minas Gerais, Brazil Campinas, São Paulo, Brazil
Image source Satellite (Google Earth Pro) UAV
Flying altitude (meters above the ground) 330 100
Date of imagery 05/16/2018 04/13/2016
Patch size (px) 1280 x 1080 240 x 240
Spatial resolution (meters per pixel) 0.15 0.03
Total area coverage (km 2 ) 31.5 3.05
N° of patches (swimming pools) 665 for training 52 for training
160 for testing 22 for testing
N° of patches (water tanks) 608 for training 102 for training
148 for testing 44 for testing
N° of annotated swimming pools 3980 65
N° of annotated water tanks 16216 196

2.2 Faster R-CNN

In this study, we used a Faster R-CNN framework for object detection, based on the work of Fernandes, Wildemberg & Dos Santos [40]. This framework is not state-of-the-art but is a relatively new and effective framework for object detection. It is widely used, and, because of that, there is substantial support material available online. The Faster R-CNN pipeline, shown in Fig 3, consists of: 1) a CNN to receive the input image and provide the feature map; 2) a Region Proposal Network (RPN) to generate bounding boxes (rectangular boxes to describe a target location) and predict the possibility of them being background or foreground; and 3) a series of fully connected layers to predict the locations of the bounding boxes in the image and their respective labels (Softmax). More details on Faster R-CNN can be seen in Ren et al. [44].

Fig 3. Faster R-CNN architecture pipeline illustration.

Fig 3

Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

2.3 Software and hardware setup

This study used Python with the DL models being replicated using Pytorch [45] version 1.40, a framework conceived to allow efficient exploitation of DL with a graphics processing unit (GPU). All experiments were performed on a product from Google Research, named Google Colab. The Colab is a free computational environment that allows anybody to write and execute python code through the browser and is especially well suited to ML, data analysis, and education. More technically, Colab is a hosted Jupyter notebook service that requires no setup to use, while providing free access to computing resources including GPU with approximately 12.72 GB of RAM. Microsoft Windows 10 Pro was used as the operating system, with ArcGIS 10.5 and Python version 3.6. Additionally, we used an open-source graphical image annotation tool called Labelme [46]. The R-4.0.3 software environment was used for statistical computing and graphics.

The UAV flight took place on April 13, 2016 on a cloudy day following a rainy day. The flight planning software used was MissionPlanner, where the overlay of the images was defined according to the recommendations of Agisoft (2016), which were 60% lateral and 80% longitudinal. The ground resolution used was approximately 0.03 m/pixel.

The flight execution stage started with verification of the ground equipment, and, since no problems were perceived, the orthomosaic of images of the base captured from 100 m height was evaluated. After this verification, the data and images that made up the mosaics were acquired from the 12.1 MP Canon PowerShot S100 (5.2 mm focal length) carried by the UAV Phantom 3 from the manufacturer DJI. The camera model used had a maximum resolution of 4000 × 3000 pixels, a focal length of 5.2 mm, 12.1 effective megapixels, a pixel pitch of 1.87 μm, a sensor size of 7.53 × 5.64 mm, and sensor resolution of 4027 × 3005 pixels. More details about the camera can be seen in Digital Camera Database [47].

We used Agisoft Photoscan 1.2.0 software for the digital processing of the data obtained by the UAV. The alignment of the images was performed through the correlation of the properly oriented overlapping images, where the algorithm searched for common points in the images and combined them, finding the position of the camera in each image and refining its calibration parameters, generating a sparse cloud of points. The five basic steps of processing with Agisoft Photoscan are: (1) automatic camera calibration based on the exchangeable image file format data of the photographs; (2) alignment of the photos from the common points between the photographs; (3) generation of the point cloud, where the x, y, and z coordinates are identified based on the estimated positions of the photographs; (4) creation of a digital model of the triangular mesh surface; and (5) generation of an orthomosaic from the texturing of the geometry constructed by the triangular mesh [48].

For the measurements, we did not use control points, and nadir images were used. For the city of Campinas, four areas were considered; in the stage of mission planning the number of projected images varied. Areas 1 to 4 had 170, 207, 458, and 314 images, respectively.

2.4 Methodology of training and prediction

The neural networks were based on the models used by Fernandes, Wildemberg & Dos Santos [40], both for water tanks and swimming pools, which used the Faster R-CNN framework and MobileNetV2 [49] as the CNN backbone. The MobileNetV2 is a light weight model with improvements focused on deep mobile computer vision applications. The training process was exactly the same as that in the cited work [40]. We used the Adam optimizer and a learning rate of 0.0001, momentum of 0.9, and weight decay of 0.0005. Then, the model was trained for 50 epochs with a batch size of 4, and a random horizontal flip (0.5) transform was used in the data loader as a form of data augmentation.

This made it possible to initialize the networks with the weights of these models trained on the dataset of Belo Horizonte [40] to speed up and improve the performance of the algorithm and then train the networks on our new dataset, adapting them to the characteristics of Campinas. This strategy is called “transfer learning”. This technique frequently used in computer vision studies where there are a limited number of images to train the network but there is a similar larger dataset that can serve as a starting point for the training process.

The global methodology for obtaining object detection was divided into six steps, which are presented in Fig 4 as a methodology flowchart for the water tank detection. From the input data (Step 1), we performed a data processing step (Step 2), because the original UAV images were larger than 15000 × 15000 pixels and over 1 GB in size, which required high processing power. To resolve this problem, we reduced the file size of the images using ArcGIS 10.5 by converting from TIFF to JPG, because the JPG is the most suitable format for training the DL model [50]. It is important to highlight that the spatial resolution remained at approximately 0.03 m/pixel, because, although the JPG files are a loss compression format, when the conversion is done properly, such as through professional software like ArcGIS, it still obtains good quality images. Additionally, each UAV image was cropped into 625 smaller patches (240 × 240-pixel clippings), because the Faster R-CNN architecture works well with such image dimensions. Thereafter, we selected the patches for the training and testing datasets (Step 3)-in this instance, only patches that had water tanks or swimming pools. It should be noted that it was not necessary to have a dataset with many samples, because the transfer learning from Belo Horizonte to Campinas showed great results. After the construction of the dataset, we manually annotated the patches using Labelme software (Step 4) [46]. After the training process, we applied the prediction procedure to all patches for the four regions, a process which consisted of realizing the object detection with the trained models for each patch in a region (Step 5) and merge all the patches to reconstruct the original image, but with the detected objects (Step 6). We set an 80% reliability threshold for the detection of an object; therefore, only good predictions were displayed in the final result of the object detection.

Fig 4. Methodology flowchart for the water tanks detection.

Fig 4

Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

We used a popular metric called mean average precision (mAP) to measure the precision of the object detector. This metric considers the ratio between true positives and the total number of predicted positives. To classify if a prediction was a true positive or a false positive an Intersection over Union (IoU) threshold was defined. The IoU measures how much the predicted bounding box overlapped with the ground truth bounding box annotation. To evaluate our object detector, we used two different IoU thresholds: 0.5 and 0.5:0.05:0.95 (the average AP for IoU threshold from 0.5 to 0.95 with a step size of 0.05). The first one is the traditional threshold for AP calculation in object detection and the latter is the primary challenge metric in the Common Objects in Context (COCO) dataset.

A work methodology flowchart is provided in S1 Fig. The flowchart illustrates the process performed to create the neural network models responsible for water tank and swimming pool detection in digital images.

2.5 Evaluating the relationship between water tanks and socioeconomic levels in the four areas of Campinas

The number of swimming pools and water tanks was obtained for each study area of Campinas, allowing the calculation of the densities of these objects per square kilometer and per 100 households. The number of these objects per study area was tested using the Chi-square test for tendency and we considered a significant result a p-value lower than 0.05. The study was conducted in the R software environment.

3. Results

3.1 Metrics

To test the networks, separate metrics were computed for water tanks installed on the roof and swimming pools, considering the four study regions of the Campinas dataset. The metrics of the two objects of study are presented in Table 2, and they were obtained with different validation conditions: trained models using only the Campinas dataset, without fine-tuning, and with fine-tuning.

Table 2. Metrics for water tanks and swimming pools.

Only Campinas Without Fine-Tuning With Fine-Tuning
AP at IoU = .50 AP at IoU = .50:.05:.95 AP at IoU = .50 AP at IoU = .50:.05:.95 AP at IoU = .50 AP at IoU = .50:.05:.95
Swimming Pools 33.81 10.48 80.27 51.65 90.23 56.9
Water Tanks 46.12 20.39 0 0 87.53 46.8

The codes of the object detection framework used, as well as the Campinas dataset and other useful data for the results presented in this study can be found at the GitHub site https://github.com/higor-sc/DL-Aedes.

When using only the Campinas dataset to train the neural network, it was clear that the metrics for swimming pools and water tanks were poor, both for an IoU threshold of 0.5 and higher IoU values, which shows that the samples were insufficient according to the quantity denoted in the column Dataset 2 of Table 1. This conclusion comes from the fact that the images in both the training and testing datasets (all from Campinas) were obtained in the exact same manner and, therefore, all have very similar features. The parameters of the models for detection were based on those indicated in Section 2.4, but we changed the learning rate to 0.005 and the batch size to 1, in addition to using 20 epochs for water tanks and 15 epochs for swimming pools.

Another result obtained refers to the validation without fine-tuning, that is, we used the models of neural networks trained in Belo Horizonte to detect swimming pools and water tanks directly. In this case, the results for swimming pools were satisfactory, with 80.27% of AP with an IoU threshold of 0.5, and when higher IoU thresholds were used the metric dropped to 51.65%; this indicates that the method detected almost all swimming pools present on an image, but the coordinates of the bounding boxes predicted were not entirely accurate. For the water tanks, the results were extremely inadequate, which shows that the Belo Horizonte and Campinas images do not have the same characteristics, and the model had difficulty making predictions. The parameters of the models for detection were the same as those indicated in Section 2.4, because we used the models trained in Belo Horizonte directly.

The last result obtained refers to the validation with fine-tuning, that is, using the models of neural networks trained in Belo Horizonte for the transfer learning and using the Campinas dataset to train the models with the characteristics of Campinas, consequently improving the results for the detection of swimming pools and water tanks through the fine-tuning technique. It is important to highlight that we used an independent subset of the Campinas dataset for the validation of the algorithm. With an IoU threshold of 0.5, we obtained 90.23% of AP for swimming pools and 87.53% of AP for water tanks, which indicates an excellent performance of the technique on the swimming pool and water tank datasets after the fine-tuning. The parameters of the models for detection were based on those indicated in Section 2.4, but we changed the learning rate to 0.005 and the batch size to 1, in addition to using 15 epochs for both the water tanks and swimming pools.

In the case without fine-tuning for the water tanks that presented an inadequate result (0% of AP), we studied the influence of the Campinas dataset on the fine-tuning process, as shown in Fig 5. The Campinas dataset used to train and validate the neural network that detects the water tanks had 146 samples in total, as listed in Table 1. To evaluate the influence of the Campinas dataset on the learning of the neural network over water tanks in Campinas, we split the dataset into percentages of the total samples, and for each percentage chosen, we trained a model neural network and evaluated its performance. The percentages chosen were: 0% (without fine-tuning in Table 2; we directly used the neural networks trained in Belo Horizonte to detect the object in Campinas), 12.5% (18 patches), 25% (37 patches), 50% (73 patches), 75% (110 patches), and 100% (146 patches; with fine-tuning in Table 2). For each percentage chosen we split 70% of the patches for training and 30% for testing, that is, in the case of using 100% of the samples (146), we used 102 samples for training and 44 samples for testing. The result indicates that, to have an interesting performance in the detection of water tanks in Campinas from the neural network trained in Belo Horizonte, it was necessary to adjust the weights of the neural network a little because, with only 18 patches of Campinas (12.5% of the total), it was already possible to obtain 68.06% of the AP with an IoU threshold of 0.5, as shown in Fig 5.

Fig 5. Influence of the Campinas dataset on the fine-tuning.

Fig 5

The higher the percentage of the Campinas dataset used, the better the performance of the neural network in terms of the AP metric. To evaluate our object detector, we used two different IoU thresholds: 0.5 (blue curve) and 0.5:0.05:0.95 (the average AP for IoU threshold from 0.5 to 0.95 with a step size of 0.05; red curve).

The four study regions of Campinas presented a certain socioeconomic disparity, both in the social sense (discussed in Section 2.1.2) and in the number of objects detected, and these may be related to vulnerability to dengue and other similar deadly diseases that can be spread through mosquitoes. In this instance, Area 4 was the poorest of the four regions, and consequently, had more exposed water tanks and fewer swimming pools. In Figs 6 and 7 it is possible to see the detection of water tanks and swimming pools.

Fig 6. Water tank detection in Area 4.

Fig 6

A sample of the UAV image without water tank detection is shown in (a) while (b) shows the water tank detection through bounding boxes. Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

Fig 7. Swimming pool detection in Area 4.

Fig 7

A sample of the UAV image without swimming pool detection is shown in (a) while (b) shows the swimming pool detection through bounding boxes. Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

Unlike the aforementioned region, Area 1 is more developed; therefore, there were more pools and fewer water tanks, as these are located inside the houses. Fig 8 shows the high number of pools.

Fig 8. Swimming pools detection in Area 1.

Fig 8

Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

The detection of water tanks and pools can be complicated due to their similar blue color; however, the trained deep neural networks use different features of the images to tell them apart. These features are not so clear because, during training, a CNN can rely on any useful property of the image to do this classification, but it is possible to list some viable features explored by the networks. The first example is the size and shape of the objects, since domestic water tanks follow a very common pattern in size and shape, usually being round and substantially smaller than a pool. We can also cite the spatial context, which is a crucial feature, due to the fact that the surroundings of a pool are very different from those of a water tank. Water tanks are found on the top of roofs or specific structures; in contrast, pools are found in external leisure areas and backyards, usually with a peculiar pattern around them. This can be used to tell them apart even if the pool is empty, dirty, or of a peculiar color.

3.2 Failures

Despite some satisfactory results, the neural network model is still susceptible to errors. This scenario can be seen in Fig 9, where detections with 80% confidence (threshold) were generated; the prediction errors are characterized by false negatives and false positives, respectively, for swimming pools and water tanks.

Fig 9. Prediction errors-false negatives and false positives.

Fig 9

In (a) there is a swimming pool that was not detected, and in (b) a blue car was detected as a water tank. Satellite images published under a CC BY license, with permission from G drones, original copyright 2016.

3.3 Relationship among swimming pools, water tanks and socioeconomic levels of the study areas

As quantitative results, we calculated the densities of swimming pools and water tanks per kilometer and 100 households in the four study areas of Campinas. Table 3 shows that, while the densities (per kilometer and 100 houses) of swimming pools decreased (except Area 2) with worsening socioeconomic levels, the densities of water tanks increased.

Table 3. Numbers and densities of swimming pools and water tanks per square kilometer and 100 households in the study areas of Campinas.

Area 1 Area 2 Area 3 Area 4
Areas Number of households 377 689 624 543
Area (km2) 0.32 0.30 0.41 0.27
Swimming pools Total UAV image 41 6 26 8
Total in the area 37 6 22 6
Swimming pools/km2 * 116.7 20.3 54.3 21.9
Swimming pools/100 households* 9.8 0.9 3.5 1.1
Water Tanks Total UAV image 8 19 39 129
Total in the area 8 18 32 99
Water tanks/km2 * 25.2 60.9 78.9 360.8
Water tanks/100 households* 2.1 2.6 5.1 18.2

* Chi-square equal to 89.159; p-value = 0.05.

Table 4 shows that the difference in the swimming pool profile is significant from the water tank profile, in the four areas, implying that the higher the socioeconomic level the higher the swimming pool densities and lower the water tank densities.

Table 4. Numbers of swimming pools and water tanks in the four study areas of Campinas.

Areas Area 1 Area 2 Area 3 Area 4 Total
Swimming pools Number 37 6 22 6 71
Line % 52.0 8.5 31.0 8.5 100.0
Water tanks Number 8 18 32 99 157
Line % 5.1 11.5 20.4 63.0 100.0
Total Number 45 24 54 105 228
Line % 19.7 10.5 23.7 46.1 100.0

4. Discussion

4.1 Swimming pool and water tank detection to infer the region’s socioeconomic status and applications

Because the number of samples were small for Campinas, our study aimed to use true-color images with high spatial resolution remote sensing sensors and DL techniques to develop an efficient procedure that allowed the identification of swimming pools and water tanks installed on the roofs in residential parcels to make a socioeconomic classification of the areas. These findings were very promising, primarily because they used relatively recent data (2016) to show the current situation in a certain area. We also found a positive relationship between areas with higher socioeconomic levels and higher densities of swimming pools, and a negative one between the same areas and higher densities of water tanks. These results have several applications in vector and disease control, showing that it is possible to detect specific targets for dengue control, and generate an updated classification in real-time for socioeconomic indexes. This is essential for the optimization and efficiency of control and surveillance programs. Our results, mainly through the use of swimming pool and water tank densities, could also have an impact in other areas of research.

The principal measure for controlling arboviruses like dengue, Zika, and chikungunya is vector control of Ae. aegypti, which is based on the elimination or treatment of breeding sites, which generally consist of containers with water found in households, such as potted plants with water, a drinking apparatus for animals, disposable containers, gutters, water tanks, and swimming pools [13, 39, 51]. This vector control is developed by the municipalities and consists of visiting the properties; however, this activity has not been sufficient to prevent dengue, Zika, and chikungunya epidemics, which are increasingly widespread. A major concern with the vector control program is the limited and temporary impact on the prevention of arboviruses cases, either because they are ineffective or because they have limited coverage [52]. Identifying areas with high levels of Ae. aegypti infestation could be an important step in identifying potential areas for the occurrence of vector-borne illnesses and could allow for the creation of more appropriate coping plans for Zika, dengue, and chikungunya fever epidemics [53].

The development of a methodology for identifying water tanks installed on roofs and swimming pools using remote sensing images and AI has the potential to optimize vector control. On one hand, this methodology would be useful for socioeconomically classifying areas [11, 12, 14, 15] to identify those at greatest risk, which, once prioritized, could make control more effective. On the other hand, eliminating breeding sites is difficult as it depends on visual identification and is often out of sight of people and in places that are difficult to access [16]. As such, it is important to identify the areas and regions most likely to develop this type of breeding site using automated methods. This would allow the public service to act directly on properties that could be considered key for the maintenance and increase of vector infestation.

Since areas of low socioeconomic levels are more prone to Ae. aegypti infestation and its associated diseases [610], investing more human and financial resources in these areas could be one way to increase the effectiveness of control activities. In general, the classification of areas with different socioeconomic levels is made using data from the demographic census, which, in Brazil, occurs every 10 years, and is often outdated. This has disadvantages like the lack of uniformity in the set of variables, changes in the boundaries of the census tracts, longer than expected periods between two surveys (e.g., the census of 2020 did not occur in Brazil), and others [12]. As shown by several authors [11, 12, 14, 15] and the results discussed herein, swimming pool and water tank densities could be used to classify areas from a socioeconomic point of view and overcome the disadvantages of using census data [12]. This classification could be a useful tool to identify risk areas for dengue and the dengue carrying mosquitoes, enabling the optimization of vector and disease control. It could also be useful to identify priority areas for investments in sanitation, housing, education, health, and other issues.

In addition to prioritizing high-risk areas, a strategy that could be used to improve the effectiveness of the Ae. aegypti control is to prioritize key properties, which contain many potential breeding grounds of immature forms of the mosquito (mainly tires, cans, and bottles, among other objects that retain water) and/or contain large breeding sites, like water tanks and swimming pools, capable of producing large amounts of mosquitoes. Sometimes, the large breeding sites are more necessary to control, because of their mosquito productivity. There may be locations with many small breeding sites, which tend to have low productivity for the winged forms of the vector, while, on the other hand, a single large-sized breeding site may present high productivity of adult mosquitoes.

Studies have presented several ways to identify key properties of a region for the presence of breeding sites [41, 5457]. The main issue with these methods is the necessity of visiting the places to classify them as key or not. Given the constant lack of human and financial resources in the health sector, there are usually not sufficient human and financial resources to identify the problem places in the entire area around Brazilian cities. Our study showed an alternative and more economical way to identify some of these key places using remote sensing images and IA.

Among the Ae. aegypti breeding sites, water tanks, and swimming pools could be considered important breeding grounds for this vector. This importance is due to the volume of water in these containers capable of producing large quantities of adult mosquitoes [58, 59] and, in general, because they are difficult to access for professionals who perform vector control activities. The water tanks placed above the roof constitute modern architecture planning and, regardless of where they are placed, could produce mosquitoes and need to be monitored [13]. In some Brazilian municipalities, household reservoirs can be recognized as the main breeding ground for Ae. aegypti [13].

Water tanks, in most areas, are located in inaccessible places, hampering inspection during routine control activities. Additionally, especially in poorer areas, they are located on slabs without a roof covering or over the roofs with the main concern being the use of inadequate tank screens, which is quite common in these areas. Even though the detection of water tanks is an important result of our study, there is a need to advance the use of remote sensing and AI to enable the identification of those not properly sealed, since they are the ones that could produce mosquitoes in greater numbers. Saini et al. [38], using Faster R-CNN for detecting overhead water tanks from satellite imagery, considered the circular pattern of the tank and respective shadow on the ground as the target object in the input image and, by transfer learning, identified the features of the Faster R-CNN Inception-V2 mode. They believed that the accuracy could be improved in the future by adding more images in a dataset with various patterns and features of the overhead water tanks.

Swimming pools are characterized as important breeding sites when they are not properly treated, which can cause substantial increase in the number of mosquitoes [16]. This occurs mainly in closed and unmaintained buildings, since these buildings are often uninhabited, mainly being up for rent or sold, and therefore the identification of the existence of non-maintained swimming pools through home visits can become practically impossible [16]. The remote sensing and AI tools used in our study were capable of detecting swimming pools that could be useful to select residences to be visited for checking the quality of their maintenance. In such cases, the ideal situation would be to identify whether the swimming pools were treated properly.

4.2 Limitations of the study

The main observed limitation of our study was the small sample number used in the training step of DL, which necessitates an enormous number of samples. Here, there were limited external water tanks on the roofs, which meant that the number of samples were not sufficient to train the neural networks. Although there are water tanks under the external roofs of any building due to the country’s water deficiency, urban external tanks on the roofs are dominant in regions with low and very low income and, in contrast, residential urban swimming pools are related to high income dwellings. Thus, the proposed methodology could have a better performance in poor districts, peripheries of big or mega cities, or small towns. For these locations, with a generally strong budget constraint, the majority of the dwellings have external water tanks on the top. Those aspects will be considered in future research.

It is also important to highlight that, the results of this study show that, even with such a small sample number, the object detection presented great accuracy with the transfer learning and fine-tuning. This is an important result, because it shows that this methodology can be reproduced and expanded to other areas, even though the sample size was small. Another important point is the fact that high-resolution UAV images, or even satellite images, are expensive, and require considerable time to be obtained, however, the cost is lower compared to that spent by the government for socioeconomic surveys and the money spent with surveillance programs. Therefore, it is important in future research to investigate the performance of publicly available images from remote sensing for the detection of water tanks and swimming pools, such as via Google Earth Pro, which was used by Fernandes, Wildemberg & Dos Santos [40] to train the Belo Horizonte dataset, or through the Google Earth Engine.

Another limitation is related to the fact that we have not conducted real inspections to check the quality of the manual classification that was made to train and validate the algorithm. This is considered a limitation because it is not easy to access all the houses in the study area to check if the water tanks and swimming pools were annotated correctly, especially water tanks, which are located in inaccessible places, like the roof, which would make it necessary to enter and access the roof in every house to verify the annotation.

4.3 Generalization of the network to other cities and sensors

One important possibility is the ability to generalize our procedures to other cities and sensors; this implies that our results for Campinas can be replicated in other regions. In our experiments, with the model trained without fine-tuning, we tried to analyze how the network behaves to generalize the model for other cities and sensors. We determined that, without fine-tuning, the results were satisfactory for swimming pools; the method detected almost all swimming pools, but the coordinates of the bounding boxes predicted were not perfectly accurate. However, for water tanks, the results were inadequate.

For the Campinas dataset, we used UAV images, while for the Belo Horizonte dataset we used satellite images. Even with this difference of sensors and locations, the network was able to generalize well; however, all the images needed to be of high resolution. We also had satellite images of the study area of Campinas, but were unable to use the images due to the resolution (0.5 m). This was because water tanks, for example, are relatively small, and different image resolutions can affect the predictions. In the case of using this neural network in other areas, it would also be necessary to adapt the dataset format to the format expected by our network for the input data. Therefore, those results show us that the network can also generalize for other sensors and cities as long as the image resolutions are high and similar to those used in our datasets and the dataset is in an appropriate format.

4.4 Detection of dirty swimming pools and open water tanks

The detection of pools and water tanks is an important step towards the optimization of the arbovirus control program, although it is the first in a series of potential developments that could be provided by the combined use of remote and DL sensing, for which there are challenges that will require new studies. Examples of these issues would be the detection of dirty swimming pools and open water tanks, which are places with a high probability of the presence of mosquitoes; the detection of objects that are not very well defined, such as asbestos roofs and slabs, which are also identified as risky places for the breeding of Ae. aegypti [18, 21]; and the use of other technologies to identify objects of interest for vector control.

When it comes to detecting specific characteristics of objects in aerial images, such as dirty swimming pools and open water tanks, resolution is a significant obstacle. For instance, even satellite images with extremely high resolution might still not be sufficient to spot those characteristic features; however, UAV imagery, like that used for this study, might just have the potential to allow for this precise classification. Moreover, even with high-resolution images available, another difficulty is obtaining the image samples of these objects. Although DL methods have the potential to successfully differentiate subtle differences among similar objects, this task requires a large amount of training data for each one of the classes, becoming prohibitive when such anomalies are rare.

4.5 Detection of other objects that are not so well delimited

It is possible to expand this work for the detection of asbestos roof tiles or roof type classification. One of the possible drawbacks of this task is that those kinds of objects are not well delineated like swimming pools and domestic water tanks, meaning that the boundaries of roofs from different buildings are not clearly defined. Higher-resolution images can eliminate some ambiguity, especially in poor and densely populated urban areas, but even with high-resolution images, the detection in this scenario is not expected to be very precise.

For such situations, where the target classes are not well delimited, a viable option is semantic segmentation [60]. In this approach, each pixel of an image is given a label, representing the class of object that it belongs to. This way, it is possible to obtain high-level contextual information about the scene represented in the input image, such as the proportions of the segmented objects and how much of the scene is occupied by them. Therefore, in the task of segmenting different types of roof tiles in a satellite image, for example, the proportion of the area with one type of roof over another can be calculated, indicating the conditions of living in that region. Wang et al. [61], for example, made use of DL techniques to perform residential roof condition assessments by segmenting different types of rooftops and artifacts in them. The authors achieved an accuracy of over 91% using the dataset assembled for the task.

4.6 Future research

The use of spectral information could be useful to identify and differentiate objects, including new possibilities for mapping urban areas globally [6266]. Other technologies that are available but are underdeveloped for detecting water tanks, swimming pools, or other objects of interest for vector control are laser sensors (such as LiDAR-Light Detection and Ranging), terrestrial and airborne technologies (drones or small aircraft), remote sensing microsatellite constellations, algorithms developed to detect small objects (like water tanks), and radar sensors (such as SAR-Synthetic Aperture Radar).

A crucial need that arises from this study is to deepen the investigations concerning the relationship between mosquito infestations and socioeconomic levels. We have a grant from the São Paulo Research Foundation (FAPESP) to continue our investigations in the city of Campinas that will enable us to evaluate the relationship between soil use and occupation, including the presence of water tanks and swimming pools and Ae. aegypti infestation. We just began to develop this research and expect to publish the first results in 2022.

5. Conclusions

Our study confirmed that, despite limitations like the small number of samples to train the algorithm, it is possible to detect water tanks installed on roofs and swimming pools using remote sensing images and DL, beyond the techniques of transfer learning. We also found a positive relationship between the swimming pool density and better socioeconomic levels and a negative relationship between the water tank density and these levels. These findings have the potential to optimize Ae. aegypti control activities, to the extent that they could help direct control actions for places and areas most at risk. The identification of these objects enables mosquito abatement personnel to schedule field inspections at specific locations for possible mosquito abatement procedures. The water tank and swimming pool densities could be used to identify areas with lower socioeconomic levels to be prioritized for the development of vector control activities. This could also have applications in other areas, such as the identification of priority areas for investments in sanitation, housing, education, and health, among other issues.

Supporting information

S1 Fig. Work methodology flowchart.

The flowchart illustrates the process performed to create the neural network models responsible for water tanks and swimming pools detection in digital images.

(JPG)

S2 Fig. Graphical abstract.

This image illustrates the steps that guide the study performed.

(JPG)

Acknowledgments

We would like to acknowledge all the collaborators who made this study possible. We would like to thank Editage (www.editage.com) for English language editing.

Abbreviations

AI

Artificial Intelligence

AP

Average Precision

CNN

Convolutional Neural Network

COCO

Common Objects in Context

DL

Deep Learning

DNN

Deep Neural Network

GPU

Graphics Processing Unit

IBGE

Brazilian Institute of Geography and Statistics

IoU

Intersection over Union

IPVS

Social Vulnerability Index of SP

ML

Machine Learning

NDSPI

Normalized Difference Swimming Pools Index

NDWI

Normalized Difference Water Index

PCA

Principal Component Analysis

RAG

Region Adjacency Graph

RF

Random Forest

RGB

Red, Green and Blue

RPN

Region Proposal Networ

SEAD

State System of Data Analysis Foundation

SVM

Support Vector Machine

UAV

Unmanned Aerial Vehicle

Data Availability

All data are available through https://github.com/higor-sc/DL-Aedes.

Funding Statement

São Paulo Research Foundation (FAPESP - https://fapesp.br) grants 2015/06687-3 (G.L.B.), 2017/10297-1 (C.L.) and 2020/01596-8 (F.C.N.). Serrapilheira Institute (https://serrapilheira.org) grant R-2011-37776 (J.A.S.). H.S.C., B.S.S., and F.C.N. were supported by grant 668/2018 from the Research Pro-Rectory of the University of São Paulo (https://prp.usp.br). National Council for Scientific and Technological Development (CNPq - https://www.gov.br/cnpq) grants 306025/2019-1 (F.C.N.), 305188/2020-8 (J.A.Q.), 311395/2018-0 (J.A.S.) and 424700/2018-2 (J.A.S.). P.F.W., E.A.M.F, and J.A.S. were supported by grant APQ-00449-17 from the Minas Gerais Research Foundation (FAPEMIG - https://fapemig.br). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Paupy C, Ollomo B, Kamgang B, Moutailler S, Rousset D, Demanou M, et al. (2010). Comparative role of Aedes albopictus and Aedes aegypti in the emergence of Dengue and Chikungunya in central Africa. Vector borne and zoonotic diseases. 2010; 10(3), 259–266. doi: 10.1089/vbz.2009.0005 [DOI] [PubMed] [Google Scholar]
  • 2.Kotsakiozi P, Gloria-Soria A, Caccone A, Evans B, Schama R, Martins AJ, et al. (2017). Tracking the return of Aedes aegypti to Brazil, the major vector of the dengue, chikungunya and Zika viruses. PLOS Neglected Tropical Diseases. 2017; 11(7), e0005653. doi: 10.1371/journal.pntd.0005653 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Chiaravalloti-Neto F. Descrição da colonização de Aedes aegypti na região de São José do Rio Preto, São Paulo. Revista da Sociedade Brasileira de Medicina Tropical. 1997; 30(4), 279–285. doi: 10.1590/s0037-86821997000400002 [DOI] [PubMed] [Google Scholar]
  • 4.OPAS/OMS Brasil [Internet]. Casos de dengue nas Américas ultrapassam 3 milhões em 2019 [cited 2020 Dec 13]. Available from: https://www.paho.org/bra/index.php?option=com_content&view=article&id=6108:casos-de-dengue-nas-americas-ultrapassam-3-milhoes-em-2019&Itemid=812
  • 5.Lorenz C, Freitas RA, Chiaravalloti-Neto F. Mayaro virus distribution in South America. Acta Tropica. 2019; 198, 105093. doi: 10.1016/j.actatropica.2019.105093 [DOI] [PubMed] [Google Scholar]
  • 6.LaDeau SL, Leisnham PT, Biehler D, Bodner D. Higher Mosquito Production in Low-Income Neighborhoods of Baltimore and Washington, DC: Understanding Ecological Drivers and Mosquito-Borne Disease Risk in Temperate Cities. Int. J. Environ. Res. Public Health. 2013; 10, 1505–1526. doi: 10.3390/ijerph10041505 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Teurlai M, Menkès CE, Cavarero V, Degallier N, Descloux E, Grangeon JP, et al. Socio-economic and climate factors associated with dengue fever spatial heterogeneity: a worked example in New Caledonia. PLOS Neglected Tropical Diseases. 2015; 9 (12):e0004211. doi: 10.1371/journal.pntd.0004211 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.de Azevedo TS, Bourke BP, Piovezan R, Sallum MAM. The influence of urban heat islands and socioeconomic factors on the spatial distribution of Aedes aegypti larval habitats. Geospatial Health. 2018; 13. doi: 10.4081/gh.2018.623 [DOI] [PubMed] [Google Scholar]
  • 9.Farinelli EC, Baquero OS, Stephan C, Chiaravalloti-Neto F. Low socioeconomic condition and the risk of dengue fever: A direct relationship. Acta Tropica. 2018; 180, 47–57. doi: 10.1016/j.actatropica.2018.01.005 [DOI] [PubMed] [Google Scholar]
  • 10.Whiteman A, Gomez C, Rovira J, Chen G, McMillan WO, Loaiza J. Aedes Mosquito Infestation in Socioeconomically Contrasting Neighborhoods of Panama City. EcoHealth. 2019; 16, 210–221. doi: 10.1007/s10393-019-01417-3 [DOI] [PubMed] [Google Scholar]
  • 11.Zambon I, Serra P, Sauri D, Carlucci M, Salvati L. Beyond the ‘Mediterranean city’: socioeconomic disparities and urban sprawl in three Southern European cities. Geografiska Annaler: Series B, Human Geography. 2017; 99(3):319–37. doi: 10.1080/04353684.2017.1294857 [DOI] [Google Scholar]
  • 12.Carlucci M, Vinci S, Lamonica GR, Salvati L. Socio-spatial Disparities and the Crisis: Swimming Pools as a Proxy of Class Segregation in Athens. Social Indicators Research. 2020. doi: 10.1007/s11205-020-02448-y [DOI] [Google Scholar]
  • 13.Lesser J, Kitron U. The Social Geography of Zika in Brazil. NACLA Report on the Americas. 2016; 48:2, 123–129. doi: 10.1080/10714839.2016.1201268 [DOI] [Google Scholar]
  • 14.Niebergall S, Löw A, Mauser W. Vulnerability in Megacities: identification of urban structures in Delhi/India using very high-resolution remote sensing images. 1st EARSeL Workshop of the SIG Urban Remote Sensing. Humboldt-Universität zu Berlin, 2–3 March 2006. [Google Scholar]
  • 15.Ayush K, Uzkent B, Burke M, Lobell D, Ermon S. Generating Interpretable Poverty Maps using Object Detection in Satellite Images. arXiv:2002.01612v2 [cs.CV] 18 Feb 2020. Available from: https://arxiv.org/pdf/2002.01612.pdf. [Google Scholar]
  • 16.McFeeters SK. Using the normalized difference water index (NDWI) within a geographic information system to detect swimming pools for mosquito abatement: A practical approach. Remote Sensing. 2013; 5(7), 3544–3561. doi: 10.3390/rs5073544 [DOI] [Google Scholar]
  • 17.Niño L. Uso de la función semivariograma y estimación kriging en el análisis espacial de un indicador entomológico de Aedes aegypti (Diptera: Culicidae). Biomédica. 2008; 28 (4). doi: 10.7705/biomedica.v28i4.64 [DOI] [PubMed] [Google Scholar]
  • 18.Sanchez L, Cortinas J, Pelaez O, Gutierrez H, Concepción D, Van Der Stuyf P. Breteau Index threshold levels indicating risk for dengue transmission in areas with low Aedes infestation. Tropical Medicine and International Health. 2010; 15(2), p. 173–175. doi: 10.1111/j.1365-3156.2009.02437.x [DOI] [PubMed] [Google Scholar]
  • 19.Landau K, Van Leeuwen W. Fine Scale Spatial Urban Land Cover Factors Associated with Adult Mosquito Abundance and Risk in Tucson, Arizona. Journal of vector ecology: journal of the Society for Vector Ecology. 2012; 37. 407–18. doi: 10.1111/j.1948-7134.2012.00245.x [DOI] [PubMed] [Google Scholar]
  • 20.Chiaravalloti-Neto F, Pereira M, Fávaro EA, Dibo MR, Mondini A, Rodrigues-Júnior AL, et al. Assessment of the relationship between entomologic indicators of Aedes aegypti and the epidemic occurrence of dengue virus 3 in a susceptible population, São José do Rio Preto, São Paulo, Brazil. Acta Tropica. 2015; 142, p. 167–177. doi: 10.1016/j.actatropica.2014.11.017 [DOI] [PubMed] [Google Scholar]
  • 21.Lorenz C, Chiaravalloti-Neto F, Lage MO, Quintanilha JA, Parra MC, Dibo MR, et al. Remote sensing for risk mapping of Aedes aegypti infestations: Is this a practical task? Acta Tropica. 2020; 205:105398. doi: 10.1016/j.actatropica.2020.105398 [DOI] [PubMed] [Google Scholar]
  • 22.Kim M, Holt JB, Eisen RJ, Padgett K, Reisen WK, Croft JB. Detection of swimming pools by geographic object-based image analysis to support West Nile Virus control efforts. Photogrammetric Engineering & Remote Sensing. 2011; 77(11), 1169–1179. [Google Scholar]
  • 23.Lowe B, Kulkarni A. Multispectral Image Analysis Using Random Forest. International Journal on Soft Computing. 2015; 6, 1–14. doi: 10.5121/ijsc.2015.6101 [DOI] [Google Scholar]
  • 24.Mountrakis G, Im J, Ogole C. Support vector machines in remote sensing: A review. ISPRS Journal of Photogrammetry and Remote Sensing. 2011; 66(3), 247–259. doi: 10.1016/j.isprsjprs.2010.11.001 [DOI] [Google Scholar]
  • 25.Goodfellow Ian, Bengio Yoshua, and Courville Aaron. Deep learning. MIT press, 2016. [Google Scholar]
  • 26.LeCun Y., Bengio Y. & Hinton G. Deep learning. Nature. 2015; 521, 436–444. doi: 10.1038/nature14539 [DOI] [PubMed] [Google Scholar]
  • 27.Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. In: Proceedings of 25th International Conference on Neural Information Processing System, NIPS; 2012, vol. 1, pp. 1097–1105. [Google Scholar]
  • 28.Li Y, Zhang H, Xue X, Jiang Y, Shen Q. Deep learning for remote sensing image classification: A survey. WIREs Data Mining Knowl Discov. 2018; 8:e1264. doi: 10.1002/widm.1264 [DOI] [Google Scholar]
  • 29.Cortes C, Vapnik V. Support vector machine. Machine learning. 1995; 20(3), 273–297. [Google Scholar]
  • 30.Tien D, Rudra T, Hope AB. Swimming pool identification from digital sensor imagery using SVM. In Proceedings—Digital Image Computing Techniques and Applications: 9th Biennial Conference of the Australian Pattern Recognition Society, DICTA 2007. 2007; 523–527. doi: 10.1109/DICTA.2007.4426841 [DOI] [Google Scholar]
  • 31.Ford A, Roberts A. Colour space conversions. Westminster University, London. 1998; 1–31. [Google Scholar]
  • 32.Kass M, Witkin A, Terzopoulos D. Snakes: Active contour models. International journal of computer vision. 1988; 1(4), 321–331. doi: 10.1007/BF00133570 [DOI] [Google Scholar]
  • 33.Galindo C, Moreno P, González J, Arevalo V. Swimming pools localization in colour high-resolution satellite images. IEEE International Geoscience and Remote Sensing Symposium. 2009; 4, 4–510. [Google Scholar]
  • 34.McFeeters SK. The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. International journal of remote sensing. 1996; 17(7), 1425–1432. doi: 10.1080/01431169608948714 [DOI] [Google Scholar]
  • 35.Saghri JA, Cary DA. A rectangular-fit classifier for synthetic aperture radar automatic target recognition. In Applications of Digital Image Processing XXX. International Society for Optics and Photonics. 2007; 6696, 66961B. doi: 10.1117/12.740236 [DOI] [Google Scholar]
  • 36.Wold S, Esbensen K, Geladi P. Principal component analysis. Chemometrics and intelligent laboratory systems. 1987; 2(1–3), 37–52. doi: 10.1016/0169-7439(87)80084-9 [DOI] [Google Scholar]
  • 37.Rodríguez-Cuenca B, Alonso MC. Semi-Automatic Detection of Swimming Pools from Aerial High-Resolution Images and LIDAR Data. Remote Sensing. 2014; 6, 2628–2646. doi: 10.3390/rs6042628 [DOI] [Google Scholar]
  • 38.Saini I, Sharnma P, Dandabathula G, Parikh D, Khandelwal S, Rao SS. Automatic detection of overhead water tanks form satellite images using Faster-RCNN. International Journal of Advanced Research in Computer Science. 2019; 10(5):34–37. doi: 10.26483/ijarcs.v10i5.6466 [DOI] [Google Scholar]
  • 39.Feria-Arroyo TP, Aguilar C, Oraby T. A tale of two cities: Aedes Mosquito surveillance across the Texas-Mexico Border. Sub Agric Envir. 2020; 71, 12–20. [Google Scholar]
  • 40.Fernandes E, Wildemberg P, dos Santos, J. Water Tanks and Swimming Pools Detection in Satellite Images: Exploiting Shallow and Deep-Based Strategies. In Anais do XVI Workshop de Visão Computacional. 2020; 117–122. Porto Alegre: SBC. doi: 10.5753/wvc.2020.13491 [DOI] [Google Scholar]
  • 41.Barbosa GL, Lage MO, Andrade VR, Gomes AHA, Quintanilha JA, Chiaravalloti-Neto F. Use of an Extended Premise Condition Index for detection of priority areas for vector control actions. Acta Tropica. 2020; 209: 105543. doi: 10.1016/j.actatropica.2020.105543 [DOI] [PubMed] [Google Scholar]
  • 42.Ferreira MP, Dini NP, Ferreira SP. Espaços e dimensões da pobreza nos municípios do Estado de São Paulo: Índice Paulista de Vulnerabilidade Social–IPVS. 2006. [Google Scholar]
  • 43.SEADE–Fundação Sistema Estadual de Análise de Dados [Internet]. Índice Paulista de Vulnerabilidade Social. Versão 2010 [cited 2020 Dec 10]. Available from: http://ipvs.seade.gov.br/view/index.php.
  • 44.Ren S, He K, Girshick R, Sun J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems—Volume 1 (NIPS’15). MIT Press, Cambridge, MA, USA. 2015; 91–99. [Google Scholar]
  • 45.Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Wallach H. M., Larochelle H., Beygelzimer A., d’Alché Buc F., Fox E. B.& Garnett R. (eds.), NeurIPS. 2019; 8024–8035. [Google Scholar]
  • 46.Wada K. Labelme: Image Polygonal Annotation with Python [cited 2020 Dec 5]. Repository: github [Internet]. Available from: https://github.com/wkentaro/labelme.
  • 47.Digital Camera Database [Internet]. Canon PowerShot S100 [cited 2021 Jul 27]. Available from: https://www.digicamdb.com/specs/canon_powershot-s100/.
  • 48.Agisoft PhotoScan User Manual: Professional Edition, Version 1.4. Copyright © 2018 Agisoft LLC, 2018 [cited 2020 Nov 20]. Web site [Internet]. Available from: https://www.agisoft.com/pdf/photoscan-pro_1_4_en.pdf.
  • 49.Sandler Mark, Howard Andrew, Zhu Menglong, Zhmoginov Andrey, and Chen Liang-Chieh. Mobilenetv2: Inverted residuals and linear bottlenecks. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2018; 4510–4520. [Google Scholar]
  • 50.Safonova A, Guirado E, Maglinets Y, Alcaraz-Segura D, Tabik S. Olive Tree Biovolume from UAV Multi-Resolution Image Segmentation with Mask R-CNN. Sensors. 2021; 21(5):1617. doi: 10.3390/s21051617 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Fouet C, Kamdem C. Integrated Mosquito Management: Is Precision Control a Luxury or Necessity? Trends in parasitology. 2019; 35(1), 85–95. doi: 10.1016/j.pt.2018.10.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Achee NL, Grieco JP, Vatandoost H, Seixas G, Pinto J, Ching-N GL, et al. Alternative strategies for mosquito-borne arbovirus control. PLOS Neglected Tropical Diseases. 2019; 13(1): e0006822. doi: 10.1371/journal.pntd.0006822 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Brady OJ, Smith DL, Scott TW, Hay SI. Dengue disease outbreak definitions are implicitly variable. Epidemics. 2015; 11, 92–102. doi: 10.1016/j.epidem.2015.03.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Tun-Lin W, Kay BH, Barnes A. The Premise Condition Index: A Tool for Streamlining Surveys of Aedes aegypti. The American Journal of Tropical Medicine and Hygiene. 1995; 53, 591–594. doi: 10.4269/ajtmh.1995.53.591 [DOI] [PubMed] [Google Scholar]
  • 55.Maciel‐de‐Freitas R, Peres RC, Souza‐Santos R, Lourenço‐de‐Oliveira R. Occurrence, productivity and spatial distribution of key-premises in two dengue endemic areas of Rio de Janeiro and their role in adult Aedes aegypti spatial infestation pattern. Tropical Medicine & International Health. 2008; 13, 1488–1494. doi: 10.1111/j.1365-3156.2008.02162.x [DOI] [PubMed] [Google Scholar]
  • 56.Manrique-Saide P, Davies CR, Coleman PG, Che-Mendoza A, Dzul-Manzanilla F, Barrera-Pérez M, et al. The Risk of Aedes aegypti Breeding and Premises Condition in South Mexico. Journal of the American Mosquito Control Association. 2013; 29, 337–345. doi: 10.2987/13-6350.1 [DOI] [PubMed] [Google Scholar]
  • 57.Peres RC, Rego R, Maciel-de-Freitas R. The use of the Premise Condition Index (PCI) to Provide Guidelines for Aedes aegypti Surveys. Journal of Vector Ecology. 2013; 38, 190–192. doi: 10.1111/j.1948-7134.2013.12027.x [DOI] [PubMed] [Google Scholar]
  • 58.Vezzani D, Velázquez SM, Schweigmann N. Seasonal pattern of abundance of Aedes aegypti (Diptera: Culicidae) in Buenos Aires city, Argentina. Memórias do Instituto Oswaldo Cruz. 2004; 99(4), 351–356. doi: 10.1590/s0074-02762004000400002 [DOI] [PubMed] [Google Scholar]
  • 59.Abe M, McCall PJ, Lenhart A, Villegas E, Kroeger A. The Buen Pastor cemetery in Trujillo, Venezuela: measuring dengue vector output from a public area. Tropical Medicine & International Health. 2005; 10(6), 597–603. doi: 10.1111/j.1365-3156.2005.01428.x [DOI] [PubMed] [Google Scholar]
  • 60.Garcia-Garcia A, Orts-Escolano S, Oprea S, Villena-Martinez V, Garcia-Rodriguez J. A review on deep learning techniques applied to semantic segmentation. arXiv preprint arXiv:1704.06857. 2017. Available from: https://arxiv.org/abs/1704.06857. [Google Scholar]
  • 61.Wang F, Kerekes JP, Xu Z, Wang Y. Residential roof condition assessment system using deep learning. Journal of Applied Remote Sensing. 2018; 12(1), 016040. doi: 10.1117/1.JRS.12.016040 [DOI] [Google Scholar]
  • 62.Swanson AL, Helmlinger MC. Spectral image classification of rooftop conditions for use in property insurance. United States—Patent Application. US 2014/0019166A1. 2014. [Google Scholar]
  • 63.Guanter L, Kaufmann H, Segl K, Foerster S, Rogass C, Chabrillat S, et al. The EnMAP Spaceborne Imaging Spectroscopy Mission for Earth Observation. Remote Sensing. 2015; 7, 8830–8857. doi: 10.3390/rs70708830 [DOI] [Google Scholar]
  • 64.Guanter L, Brell M, Chan JCW, Giardino C, Gomez-Dans J, Mielke C, et al. Synergies of Spaceborne Imaging Spectroscopy with Other Remote Sensing Approaches. Surveys in Geophysics. 2019; 40, 657–87. doi: 10.1007/s10712-018-9485-z [DOI] [Google Scholar]
  • 65.Linden S, Okujeni A, Canters F, Degerickx J, Heiden J, Hostert P, et al. Imaging Spectroscopy of Urban Environments. Surveys in Geophysics. 2019; 40(3): 471–88. doi: 10.1007/s10712-018-9486-y [DOI] [Google Scholar]
  • 66.Braun A, Warth G, Bachofer F, Hochschild V. Identification of roof materials in high-resolution multispectral images for urban planning and monitoring. Joint Urban Remote Sensing Event (JURSE), Vannes, France. 2019; 1–4, doi: 10.1109/JURSE.2019.8809026 [DOI] [Google Scholar]

Decision Letter 0

Guy J-P Schumann

29 Jun 2021

PONE-D-21-10208

Water tank and swimming pool detection based on remote sensing and deep learning: Relationship with socioeconomic level and applications in dengue control

PLOS ONE

Dear Dr. Cunha,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

This paper is interesting but many lacks detailed explanations on several methodological aspects as highlighted by both reviewers. I therefore encourage the authors in their revision to respond to each comment carefully and focus on detailing the methodology better.

Please submit your revised manuscript by Aug 13 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Guy J-P. Schumann

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Methods section, please provide additional information regarding the permits you obtained for the work. Please ensure you have included the full name of the authority that approved the field site access and, if no permits were required, a brief statement explaining why.

3. In your Methods section, please provide additional location information, including geographic coordinates for the data set if available.

4. We note that Figures 3, 4, 6, 7, 8 and 9 in your submission contain satellite images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth). For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

We require you to either (a) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (b) remove the figures from your submission:

a) You may seek permission from the original copyright holder of Figure(s) [#] to publish the content specifically under the CC BY 4.0 license.

We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

b) If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

The following resources for replacing copyrighted map figures may be helpful:

USGS National Map Viewer (public domain): http://viewer.nationalmap.gov/viewer/

The Gateway to Astronaut Photography of Earth (public domain): http://eol.jsc.nasa.gov/sseop/clickmap/

Maps at the CIA (public domain): https://www.cia.gov/library/publications/the-world-factbook/index.html and https://www.cia.gov/library/publications/cia-maps-publications/index.html

NASA Earth Observatory (public domain): http://earthobservatory.nasa.gov/

Landsat: http://landsat.visibleearth.nasa.gov/

USGS EROS (Earth Resources Observatory and Science (EROS) Center) (public domain): http://eros.usgs.gov/#

Natural Earth (public domain): http://www.naturalearthdata.com/

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Summary of Research and Overall Impression

The authors sought to use true-color digital imagery, obtained by using a UAV, to determine if open water tanks, that were located on rooftops, could be detected and differentiated from swimming pools, and use this information to assist public health programs in the treatment of vector-borne diseases. They made use of annotated Google Earth imagery of one region to create training datasets that were used to improve the outcome of the analysis of data acquired by the UAV’s.

My overall impression is that it’s a good study. I think that the approach is an efficient one that is low-cost, and which could provide timely information regarding the potential for the spreading of vector-borne diseases in an urban setting.

Discussion of Specific Areas of Improvement

Major Issues

I think that a major question is “Why are the water tanks even on the buildings?” Presumably, they collect rainfall for consumptive purposes, but their presence is not explained. Perhaps the occupants must use them to collect all of the water that they use. Perhaps these buildings have water lines, but the occupants must pay too much money for the water and the tanks help offset the cost. The presence or the absence of water tanks must be made relevant to the reader. Also, the relevance of the presence, or absence, of swimming pools should also be made clear to the reader. Are there high costs to build and to maintain the pools (e.g. high cost of water)? Are there taxes which must be paid to have the swimming pools? The authors make it clear that water tanks and swimming pools need to be detected, but I don’t think that they sufficiently explain the relevance as to why.

Sec 1.2 Line 91: The authors state “Many studies have applied machine learning (ML) to identify soil cover.” and then they cite just two references. I would propose that “many studies” would mean more than just two studies. I think that more studies should be cited, because perhaps two are not enough, but the authors don’t need an excessive number to support their approach. The sentence should also be rewritten to support the increase in citations.

Sec 1.3 Line 99: The authors state that “A major motivation for this field is the automation of tasks that require analyzing substantial amounts of visual data in a short period of time.” but they cite no reference to support their assertion.

Line 187-188: The flying altitude of the Google Earth was 330 meters, but it was not indicated if this represented elevation above the ground, or above mean sea-level. This should be made clear to the reader.

Line 262: Inconsistent units were used for linear measurements related to the camera on the UAV. Focal length was reported in millimeters, but ground resolution was reported as 3 cm/pixel (my own calculation was 0.35769 meters, or 35.8 millimeters). That said, Line 262 states that the pixel size was 1.86 by 1.86 μm (micrometers), and this is not possible. A pixel size of that magnitude would be in the short-infrared wavelengths. This pixel size is about 1/19,000 to small on a linear measure, or about 1/316,000,000 too small. The pixel size should be recalculated, and also a more detailed description of how these numbers were determined would be helpful to the reader. Perhaps, an equation could be inserted to show how the calculations were performed.

Line 286: The “quality” of the images was decreased, and file extensions were converted from .tiff to .jpg. This is a major issue. I presume that the spatial resolution was degraded, but what was the original spatial resolution and what was the degraded spatial resolution? Also, what effect did the conversion of a .tiff file format into a .jpeg file format have upon the integrity of the original data? TIFF files are lossless, and maintain the original integrity of the data, but JPEG files are, most of the time, a lossy compression format that does not maintain the original integrity of the data.

Section 2.1.1 and 2.1.2: The climatological data for the data for the two sites are inconsistent with one another, and they appear to lack any citation.

Line 214-215: The authors introduce the term “Social Vulnerability Index of SP (IPVS)”, but do not explain what it represents. They do mention that some areas have “exceptionally low vulnerability” while others have “exceptionally low, low, and medium vulnerability”, and another area has “mostly extremely high vulnerability”, but the authors do not explain what the vulnerabilities represent. I think that a thorough explanation of the terms would improve the readability of the text by non-Brazilians.

Related to the previous issue: Line 368: The authors suggest that “the socioeconomic level is related to the level of risk in the region.” I think that it might be helpful to remind the reader as to what is represented by “the level of risk”.

Minor Issues

Line 381 “…greater number of pools and few water tanks…”. I think that the word “few” should actually be “fewer” to improve the readability of the line.

Line 415 (and perhaps others): I think that the term “non-multispectral” could be confused with “panchromatic”. The term multispectral is typically associated with visible and near-infrared, and possibly even shortwave infrared, wavelengths. It is clear that up to this point, the authors are using true-color images which are typically related to spectral wavelengths from 0.4 to 0.7 micrometers. I would recommend that the term “non-multispectral” be replaced with the term “true-color” as that was the spectral region used in the study.

Line 489-490: The term “maintenance-free” implies that the buildings need not be maintained, however, I believe that the authors meant to imply that proper maintenance was not being conducted and so I would suggest the word “unmaintained” instead of “maintenance-free”.

Section 4.6 should be “Future research”

It would be helpful for the list of abbreviations to be alphabetized.

The quality of the language is adequate, but as a native English speaker, I would have written some statements differently. The authors used Editage (editage.com) for the English language editing, but I’m not certain that a native English speaker actually performed the editing.

Reviewer #2: This study aimed to detect water tanks installed on rooftops and swimming pools in digital images for the identification and classification of areas based on the socioeconomic index, to assist public health programs in the control of diseases linked to the Aedes aegypti mosquito. The manuscript is well written and easy to read. However, it needs to be supplied with sufficient details to allow its reproducibility. On the other hand, some methodological aspects also need to be clarified to improve the quality of the manuscript, therefore, the following comments:

Please include the spatial resolution (not image size/patch size) of the images (Google Earth Pro and UAV) on the Table 1 since it is an important information for the reproducibility of the study.

It has been mentioned that the UAV flight took place on April 13, 2016. However, the weather conditions in that day are not described. Taking into consideration that images are captured by a camera lenses with an optical sensor, how can the ambient condition (e.g., rainy, sunny, cloudy, etc.) at the time of the image capturing affect the image properties, thus affecting Deep Learning (DL) model results? Also, what was the configuration of the camera prior the flight? e.g., ISO, White Balance, Shutter Speed, etc.?

The Authors said that model with and without fine-tuning were implemented. However, the description of respective model parameters is not clear.

On the methodology, the description of probable image properties /features used to train the DL model is missing. Given that most of the swimming pools and water tanks exhibited on the figures are of blue color, which particular features/properties are used to differentiate both? Is the size of the object or the number of pixels or the shape type considered? Is there any specific RGB color range adopted for each object? If for example an swimming pool is temporary empty or dirty in the day of the survey, can it be detected with your approach? Such information is not clear in the current manuscript.

Theoretically, for example, two swimming pools with different colors of the finishing material (e.g., white ceramic tiles and green ceramic tiles), are expected to have also different colors when are filled with water. Given that field visits were not made, how is the presented approach capable of correctly labeling the objects (swimming pools and water tanks) prior to DL model? I am worried if the (probably) optical recognition and manual identification and labeling (Lines 292 – 293) may miss recognize some swimming pools and water tanks thus misleading the model. How can this approach secure that the object labeling before model implementation is done accurately?

If not mistaken, most of the water tanks in the presented figures are of blue color (e.g., Fig 6, Fig 7, Fig 9). All water tanks in the study area are of the same color? Is it possible to give more detailed descriptions about the characteristic (e.g., dimensions, shape) of target objects in the study area? Furthermore, is this kind of object characteristics took into account in your DL model?

It has been stated that Campinas dataset samples were insufficient, thus results were poor (Lines 325 – 327). Having more samples (probably with same/similar characteristics as the actual samples) would improve the DL model performance? I wonder if the samples themselves on the training set for swimming pools and water tanks did not have the same/similar characteristic with those on the test set. Meaning that the model was constructed (trained) with sample features that were not able to render the sample features on the test set. If this is the case, poor model performance can be expected in any data driven model. Don’t the Authors think this to be only a qualitative problem but not quantitative problem as suggested?

Although areas with low socioeconomic level are often linked to high infestation of mosquitos, (Line 446 – 447), is there any evidence from your study supporting those areas with high density of water tanks are more prone to Ae. aegypti infestation than those with high density of swimming pools? One of the most notorious downsides of this study is that no field inspection to collect and compare the density of mosquitos between the two classified areas was done. Such information would eventually offer a scientific evidence for the validation of the model.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Dec 9;16(12):e0258681. doi: 10.1371/journal.pone.0258681.r002

Author response to Decision Letter 0


13 Aug 2021

RESPONSE TO REVIEWERS

August 2021

Dear Dr. Guy J-P. Schumann

Academic Editor

PLOS ONE

We are resubmitting the manuscript entitled “Water tank and swimming pool detection based on remote sensing and deep learning: Relationship with socioeconomic level and applications in dengue control”.

Thank you for your kind reply. We also thank the reviewers’ suggestions and comments. They were very important for the improvement of our manuscript. Please find below our answers to the reviewers’ comments and the list of changes in the manuscript. All changes in the manuscript were highlighted with Track Changes.

We forwarded the new version of the manuscript to the Editage (editage.com) and the editor revised the English language again to make the language quality adequate. In the submission process we will upload the “Editing Certificate” provided by Editage.

We look forward to hearing from you soon.

Sincerely,

Higor Souza Cunha

---------------------------------------------------------------

I – ANSWERS TO ACADEMIC EDITOR

Dear Academic Editor,

The authors are very grateful for your comments. Please check below our answers.

Point 1: Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming.

Answer: We ensure that the manuscript follows the PLOS ONE’s style requirements.

Point 2: In your Methods section, please provide additional information regarding the permits you obtained for the work. Please ensure you have included the full name of the authority that approved the field site access and, if no permits were required, a brief statement explaining why.

Answer: In the Material and Methods we provided the additional information regarding the permits obtained for the work. We added the following text on lines 241 to 243 (see Track Changes version):

“The study on the four areas described above required the approval of the Ethics Committee in Plataforma Brasil (Brazil Platform System), according to the presentation certificate for ethical evaluation no. 43813015.9.0000.0059 and approved according to opinion no. 1.082.780.”

Point 3: In your Methods section, please provide additional location information, including geographic coordinates for the data set if available.

Answer: In the Section “Material and methods” we corrected the geographic coordinates of Campinas in the first paragraph of Section 2.1.2 (Campinas) and included the geographic coordinates of the four areas of Campinas in the second paragraph of this item. We also provide the geographic coordinates of the areas of Belo Horizonte (Section 2.1.1). The datum adopted was the WGS84.

Point 4: 4. We note that Figures 3, 4, 6, 7, 8 and 9 in your submission contain satellite images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth).

We require you to either (a) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (b) remove the figures from your submission:

Answer: We contact the original copyright holder and in the submission process we will upload the completed “Content Permission Form” under the CC BY 4.0 License.

We are now permitted to publish the figures contain satellite images obtained by the company G drones (http://www.g-drones.com.br/) under a CC BY license, with permission from G drones, original copyright 2016. The written permission granted by George Alfredo Longhitano, director of the G drones.

This way, the Figures 3, 4, 6, 7, 8, and 9 will comply with the PLOS ONE guidelines.

---------------------------------------------------------------

II – ANSWERS TO REVIEWER 1

Dear Reviewer 1,

The authors are very grateful for your comments. Please check below our answers.

The authors sought to use true-color digital imagery, obtained by using a UAV, to determine if open water tanks, that were located on rooftops, could be detected and differentiated from swimming pools, and use this information to assist public health programs in the treatment of vector-borne diseases. They made use of annotated Google Earth imagery of one region to create training datasets that were used to improve the outcome of the analysis of data acquired by the UAV’s.

My overall impression is that it’s a good study. I think that the approach is an efficient one that is low-cost, and which could provide timely information regarding the potential for the spreading of vector-borne diseases in an urban setting.

MAJOR ISSUES

Point 1: I think that a major question is “Why are the water tanks even on the buildings?” Presumably, they collect rainfall for consumptive purposes, but their presence is not explained. Perhaps the occupants must use them to collect all of the water that they use. Perhaps these buildings have water lines, but the occupants must pay too much money for the water and the tanks help offset the cost. The presence or the absence of water tanks must be made relevant to the reader. Also, the relevance of the presence, or absence, of swimming pools should also be made clear to the reader. Are there high costs to build and to maintain the pools (e.g. high cost of water)? Are there taxes which must be paid to have the swimming pools? The authors make it clear that water tanks and swimming pools need to be detected, but I don’t think that they sufficiently explain the relevance as to why.

Answer: In Brazil, especially in regions with low socioeconomic level, the water tanks are often located on the top of the roof of homes, exposed to the weather.

In the Section “1.1 Motivations to develop the study”, we assume as a hypothesis, based on the literature, that water tanks and swimming pools are useful objects in the classification of areas according to socioeconomic levels, principally areas with poor socioeconomic levels and that are conducive to the development of Aedes aegypti mosquito. Given this premise, the presence or absence of water tanks and swimming pools it is an important factor, because a higher density of swimming pools could be a good indicator of areas with better socioeconomic levels, as well as a higher density of water tanks installed on the roof could be a good indicator of areas less privileged.

The identification of areas at greatest risk of the presence of the vector and the occurrence of the diseases is one of the measures adopted to optimize vector and disease control. The water tanks and swimming pools could be characterized as key breeding sites. The main issue with water tanks is the inadequate use of tank screens, allowing vector access and breeding. As for swimming pools, they could become significant breeding grounds for mosquitoes when they are not properly treated. The objects were chosen based on one of the most important elements in the life cycle of this species of mosquito, i.e., the water.

In Section 4.6 Future research, we emphasize that our investigation in the city of Campinas continue and that new technologies are being developed to relate the presence of water tanks and swimming pools with the infestation of Ae. aegypti.

Point 2: Sec 1.2 Line 91: The authors state “Many studies have applied machine learning (ML) to identify soil cover.” and then they cite just two references. I would propose that “many studies” would mean more than just two studies. I think that more studies should be cited, because perhaps two are not enough, but the authors don’t need an excessive number to support their approach. The sentence should also be rewritten to support the increase in citations.

Answer: Thank you for pointing out this issue. We replaced the word “many” with “some”, as the intention was just to contextualize the application to the reader. In the cited references, the reader can find other studies in the same scope. Such change is shown on line 92 (see Track Changes version).

Point 3: Sec 1.3 Line 99: The authors state that “A major motivation for this field is the automation of tasks that require analyzing substantial amounts of visual data in a short period of time.” but they cite no reference to support their assertion.

Answer: Thank you for pointing out this issue. We added a reference that supports the assertion (line 103, see Track Changes version).

Point 4: Line 187-188: The flying altitude of the Google Earth was 330 meters, but it was not indicated if this represented elevation above the ground, or above mean sea-level. This should be made clear to the reader.

Answer: Thank you for pointing out this issue. The flying altitude is referenced to elevation above the ground. We provide this information in the Table 1 – “Comparison between the Belo Horizonte and Campinas datasets” (see Track Changes version).

Point 5: Line 262: Inconsistent units were used for linear measurements related to the camera on the UAV. Focal length was reported in millimeters, but ground resolution was reported as 3 cm/pixel (my own calculation was 0.35769 meters, or 35.8 millimeters). That said, Line 262 states that the pixel size was 1.86 by 1.86 μm (micrometers), and this is not possible. A pixel size of that magnitude would be in the short-infrared wavelengths. This pixel size is about 1/19,000 to small on a linear measure, or about 1/316,000,000 too small. The pixel size should be recalculated, and also a more detailed description of how these numbers were determined would be helpful to the reader. Perhaps, an equation could be inserted to show how the calculations were performed.

Answer: We confirmed the technical specifications of the camera and in fact it has a pixel pitch of 1.87 μm, with sensor size of 7.53 x 5.64 mm. We made the following change to the lines 289 to 292 (see Track Changes version):

“The camera model used had a maximum resolution of 4000 ×x 3000 pixels, a focal length of 5.2 mm, 12.1 effective megapixels, and a pixel size pitch of 1.86 x 1.86 1.87 µm, a sensor size of 7.53 × 5.64 mm, and sensor resolution of 4027 × 3005 pixels. More details about the camera can be seen in Digital Camera Database [47].”

We chose to indicate the reference of the camera supplier, so the reader will be able to have a better understanding of the technical data of the equipment. Detailing the calculation of parameters would require the explanation of technical concepts and this is already done in the reference indicated.

Regarding the units, we think it is coherent to follow the standard established in the camera's technical manuals, as this would facilitate the purchase of the camera by a possible reader interested in the application.

Point 6: Line 286: The “quality” of the images was decreased, and file extensions were converted from .tiff to .jpg. This is a major issue. I presume that the spatial resolution was degraded, but what was the original spatial resolution and what was the degraded spatial resolution? Also, what effect did the conversion of a .tiff file format into a .jpeg file format have upon the integrity of the original data? TIFF files are lossless, and maintain the original integrity of the data, but JPEG files are, most of the time, a lossy compression format that does not maintain the original integrity of the data.

Answer: We improved the explanation of the procedure from the following text change (lines 324 to 329, see Track Changes version):

“To resolve this problem, we reduced the file size of the images using ArcGIS 10.5 by converting from TIFF to JPG, because the JPG is the most suitable format for training the DL model [50]. It is important to highlight that the spatial resolution remained at approximately 0.03 m/pixel, because, although the JPG files are a loss compression format, when the conversion is done properly, such as through professional software like ArcGIS, it still obtains good quality images.”

Point 7: Section 2.1.1 and 2.1.2: The climatological data for the data for the two sites are inconsistent with one another, and they appear to lack any citation.

Answer: We added the climatological data of Belo Horizonte (lines 189 to 190, see Track Changes version) and it is important to highlight the difference in relation to Campinas. Despite being regions of the same country, Brazil, their climates are different.

Point 8: Line 214-215: The authors introduce the term “Social Vulnerability Index of SP (IPVS)”, but do not explain what it represents. They do mention that some areas have “exceptionally low vulnerability” while others have “exceptionally low, low, and medium vulnerability”, and another area has “mostly extremely high vulnerability”, but the authors do not explain what the vulnerabilities represent. I think that a thorough explanation of the terms would improve the readability of the text by non-Brazilians.

Answer: We agree with the reviewer that the IPVS presentation needed more details. We rewrote the second paragraph of Section 2.1.2 Campinas to deal with this issue. We explained, in this new version of the manuscript, the meaning of IPVS and of each one of the levels of vulnerability (lines 228 to 240, see Track Changes version).

Point 9: Related to the previous issue: Line 368: The authors suggest that “the socioeconomic level is related to the level of risk in the region.” I think that it might be helpful to remind the reader as to what is represented by “the level of risk”.

Answer: Thank you for pointing out this issue. We changed the phrase to improve the explanation and added the following text on lines 426 to 429 (see Track Changes version):

“The four study regions of Campinas presented a certain socioeconomic disparity, both in the social sense (discussed in Section 2.1.2) and in the number of objects detected, and these may be related to vulnerability to dengue and other similar deadly diseases that can be spread through mosquitoes.”

MINOR ISSUES ISSUES

Point 1: Line 381 “…greater number of pools and few water tanks…”. I think that the word “few” should actually be “fewer” to improve the readability of the line.

Answer: We accepted this suggestion. Such change is present on line 444 (see Track Changes version).

Point 2: Line 415 (and perhaps others): I think that the term “non-multispectral” could be confused with “panchromatic”. The term multispectral is typically associated with visible and near-infrared, and possibly even shortwave infrared, wavelengths. It is clear that up to this point, the authors are using true-color images which are typically related to spectral wavelengths from 0.4 to 0.7 micrometers. I would recommend that the term “non-multispectral” be replaced with the term “true-color” as that was the spectral region used in the study.

Answer: We accepted this suggestion. Such change is present on line 493 to 494 (see Track Changes version).

Point 3: Line 489-490: The term “maintenance-free” implies that the buildings need not be maintained, however, I believe that the authors meant to imply that proper maintenance was not being conducted and so I would suggest the word “unmaintained” instead of “maintenance-free”.

Answer: We accepted this suggestion. Such change is present on lines 567 to 568 (see Track Changes version).

Point 4: Section 4.6 should be “Future research”

Answer: We corrected it.

Point 5: It would be helpful for the list of abbreviations to be alphabetized.

Answer: We corrected it.

Point 6: The quality of the language is adequate, but as a native English speaker, I would have written some statements differently. The authors used Editage (editage.com) for the English language editing, but I’m not certain that a native English speaker actually performed the editing.

Answer: Thank you for pointing out this issue. We forwarded the new version of the manuscript to the Editage (editage.com) and the editor revised the English language again to make the language quality adequate. In the submission process we will upload the “Editing Certificate” provided by Editage.

---------------------------------------------------------------

III – ANSWERS TO REVIEWER 2

Dear Reviewer 2,

The authors are very grateful for your comments. Please check below our answers.

This study aimed to detect water tanks installed on rooftops and swimming pools in digital images for the identification and classification of areas based on the socioeconomic index, to assist public health programs in the control of diseases linked to the Aedes aegypti mosquito. The manuscript is well written and easy to read. However, it needs to be supplied with sufficient details to allow its reproducibility. On the other hand, some methodological aspects also need to be clarified to improve the quality of the manuscript, therefore, the following comments:

Point 1: Please include the spatial resolution (not image size/patch size) of the images (Google Earth Pro and UAV) on the Table 1 since it is an important information for the reproducibility of the study.

Answer: We included the spatial resolution of the images in the Table 1, whose unit is meters per pixel, in the case 0.15 for Google Earth Pro images and 0.03 for UAV images.

Point 2: It has been mentioned that the UAV flight took place on April 13, 2016. However, the weather conditions in that day are not described. Taking into consideration that images are captured by a camera lenses with an optical sensor, how can the ambient condition (e.g., rainy, sunny, cloudy, etc.) at the time of the image capturing affect the image properties, thus affecting Deep Learning (DL) model results? Also, what was the configuration of the camera prior the flight? e.g., ISO, White Balance, Shutter Speed, etc.?

Answer: The images were captured on a cloudy day following a rainy day. Such context increases the discrimination of green targets/vegetation and potential breeding grounds of Ae. aegypti when using optical sensors, in addition to reducing the confusion related to the presence of clouds. These details involving environmental conditions improved the DL model results, because the detected objects (water tanks and swimming pools) were better discriminated.

We included the description of the weather conditions on line 282 (see Track Changes version).

The UAV images were obtained from a photogrammetric flight company, in this case, a mosaic of images, so we do not have more specific information about the configuration of the camera prior the flight.

Point 3: The Authors said that model with and without fine-tuning were implemented. However, the description of respective model parameters is not clear.

Answer: Thank you for pointing out this issue. We improved throughout Section 3.1 Metrics the explanation about the parameters of the models used.

Point 4: On the methodology, the description of probable image properties /features used to train the DL model is missing. Given that most of the swimming pools and water tanks exhibited on the figures are of blue color, which particular features/properties are used to differentiate both? Is the size of the object or the number of pixels or the shape type considered? Is there any specific RGB color range adopted for each object? If for example an swimming pool is temporary empty or dirty in the day of the survey, can it be detected with your approach? Such information is not clear in the current manuscript.

Answer: We added such information in the Section 3 Results, specifically in 3.1 Metrics, because the visual results through the figures illustrate the context (lines 451 to 460, see Track Changes version).

The first example is the size and shape of the objects, since domestic water tanks follow a very common pattern in size and shape, usually being round and substantially smaller than a pool. We can also cite the spatial context, which is a crucial feature, due to the fact that the surroundings of a pool are very different from those of a water tank. Water tanks are found on the top of roofs or specific structures; in contrast, pools are found in external leisure areas and backyards, usually with a peculiar pattern around them. This can be used to tell them apart even if the pool is empty, dirty, or of a peculiar color.

Point 5: Theoretically, for example, two swimming pools with different colors of the finishing material (e.g., white ceramic tiles and green ceramic tiles), are expected to have also different colors when are filled with water. Given that field visits were not made, how is the presented approach capable of correctly labeling the objects (swimming pools and water tanks) prior to DL model? I am worried if the (probably) optical recognition and manual identification and labeling (Lines 292 – 293) may miss recognize some swimming pools and water tanks thus misleading the model. How can this approach secure that the object labeling before model implementation is done accurately?

Answer: Thank you for pointing out this issue. We made a manual classification, all the water tanks and swimming pools were classified by humans, but we did not make the real inspections. As those objects are easy to visually identify in UAV images, so we assumed that the quality of the manual classification was good. Thus, we understand the fact of not having done real inspections for checking the water tanks is a limitation of our work, especially because it is not easy to have access to all the houses in the area of the images. In the Section 4.2 Limitations of the study, we explained such scenario.

It is important to highlight that the neural network model developed for Belo Horizonte improves the context of this limitation. This is because the neural network model was built with a high number of samples and images with a size of 3840x2160 pixels (4K), so the human recognition of objects was even easier due to the quality of the images. As the neural network model developed for Campinas is based on the Belo Horizonte model, such limitation was not a problem.

Point 6: If not mistaken, most of the water tanks in the presented figures are of blue color (e.g., Fig 6, Fig 7, Fig 9). All water tanks in the study area are of the same color? Is it possible to give more detailed descriptions about the characteristic (e.g., dimensions, shape) of target objects in the study area? Furthermore, is this kind of object characteristics took into account in your DL model?

Answer: Thank you for pointing out this issue. In Brazil context, water tanks are mostly blue, and we have lots of asbestos water tanks, so we had to train the algorithm according to this context.

In the case this study is extended to other countries in the future, it would be necessary to have more samples to adapt to this new context, to identify the object deeply in terms of color, texture, and shape, contemplating other types of water tanks. The same goes for swimming pools.

About the characteristics of target objects in the study area and how the model DL takes these characteristics into account, we added such information in the Section 3 Results, specifically in 3.1 Metrics, because the visual results through the figures illustrate the context (lines 451 to 460, see Track Changes version).

Point 7: It has been stated that Campinas dataset samples were insufficient, thus results were poor (Lines 325 – 327). Having more samples (probably with same/similar characteristics as the actual samples) would improve the DL model performance? I wonder if the samples themselves on the training set for swimming pools and water tanks did not have the same/similar characteristic with those on the test set. Meaning that the model was constructed (trained) with sample features that were not able to render the sample features on the test set. If this is the case, poor model performance can be expected in any data driven model. Don’t the Authors think this to be only a qualitative problem but not quantitative problem as suggested?

Answer: Having more samples would definitely improve the performance of the DL model, because in the particular experiment done, we took into account only Campinas dataset to train the neural network. For such an application to have interesting results, a high number of samples is needed, as in the case of the Belo Horizonte dataset.

The samples present in the training and test datasets have the same quality, as they were obtained in exactly the same way, so the result leads us to believe that this is a quantitative problem.

Point 8: Although areas with low socioeconomic level are often linked to high infestation of mosquitos, (Line 446 – 447), is there any evidence from your study supporting those areas with high density of water tanks are more prone to Ae. aegypti infestation than those with high density of swimming pools? One of the most notorious downsides of this study is that no field inspection to collect and compare the density of mosquitos between the two classified areas was done. Such information would eventually offer a scientific evidence for the validation of the model.

Answer: We totally agree with the reviewer and we are now developing studies to deepen the investigations about the relationship between the mosquito infestation and socioeconomic level. We have a grant from the São Paulo Research Foundation (FAPESP) to continue our investigations in the city of Campinas that will capable us to evaluate the relationship between the soil use and occupation, including the presence of water tanks and swimming pools, and Ae. aegypti infestation. We just began to develop this research and we will probably have the first results in 2022. We included a new paragraph at the end of item 4.6 Future research explaining this issue (lines 660 to 664, see Track Changes version).

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Guy J-P Schumann

4 Oct 2021

Water tank and swimming pool detection based on remote sensing and deep learning: Relationship with socioeconomic level and applications in dengue control

PONE-D-21-10208R1

Dear Dr. Cunha,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Guy J-P. Schumann

Section Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I believe that all of my concerns have been addressed. I don't have any experience with deep learning techniques, and so I did not comment with regard to the rigor of any statistical analyses.

I found the revision to be much improved over the initial submission, and the authors are to be commended.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Acceptance letter

Guy J-P Schumann

1 Dec 2021

PONE-D-21-10208R1

Water tank and swimming pool detection based on remote sensing and deep learning: Relationship with socioeconomic level and applications in dengue control

Dear Dr. Cunha:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Guy J-P. Schumann

Section Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Work methodology flowchart.

    The flowchart illustrates the process performed to create the neural network models responsible for water tanks and swimming pools detection in digital images.

    (JPG)

    S2 Fig. Graphical abstract.

    This image illustrates the steps that guide the study performed.

    (JPG)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    All data are available through https://github.com/higor-sc/DL-Aedes.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES