Skip to main content
PLOS One logoLink to PLOS One
. 2021 Oct 20;16(10):e0248797. doi: 10.1371/journal.pone.0248797

Bioacoustics for in situ validation of species distribution modelling: An example with bats in Brazil

Frederico Hintze 1,2,*, Ricardo B Machado 3, Enrico Bernard 1
Editor: Daniel de Paiva Silva4
PMCID: PMC8528307  PMID: 34669707

Abstract

Species distribution modelling (SDM) gained importance on biodiversity distribution and conservation studies worldwide, including prioritizing areas for public policies and international treaties. Useful for large-scale approaches and species distribution estimates, it is a plus considering that a minor fraction of the planet is adequately sampled. However, minimizing errors is challenging, but essential, considering the uses and consequences of such models. In situ validation of the SDM outputs should be a key-step—in some cases, urgent. Bioacoustics can be used to validate and refine those outputs, especially if the focal species’ vocalizations are conspicuous and species-specific. This is the case of echolocating bats. Here, we used extensive acoustic monitoring (>120 validation points over an area of >758,000 km2, and producing >300,000 sound files) to validate MaxEnt outputs for six neotropical bat species in a poorly-sampled region of Brazil. Based on in situ validation, we evaluated four threshold-dependent theoretical evaluation metrics’ ability in predicting models’ performance. We also assessed the performance of three widely used thresholds to convert continuous SDMs into presence/absence maps. We demonstrated that MaxEnt produces very different outputs, requiring a careful choice on thresholds and modeling parameters. Although all theoretical evaluation metrics studied were positively correlated with accuracy, we empirically demonstrated that metrics based on specificity-sensitivity and sensitivity-precision are better for testing models, considering that most SDMs are based on unbalanced data. Without independent field validation, we found that using an arbitrary threshold for modelling can be a precarious approach with many possible outcomes, even after getting good evaluation scores. Bioacoustics proved to be important for validating SDMs for the six bat species analyzed, allowing a better refinement of SDMs in large and under-sampled regions, with relatively low sampling effort. Regardless of the species assessing method used, our research highlighted the vital necessity of in situ validation for SDMs.

Introduction

Species distribution modelling (SDM) gained importance worldwide in the development of studies on biodiversity distribution and conservation [1, 2]. These distribution models can be produced and refined by crossing species presence records with biological and non-biological variables and environmental data [e.g., 3]. This is an advantage for large-scale approaches and species distribution estimates since few parts of our planet have been adequately sampled [4, 5]. Moreover, SDM has been used to support decision-making processes, including prioritizing areas and regions in public policies and international treaties [2, 5].

For biodiversity conservation purposes, imprecise models can undermine the calculation/estimate of a species’ occupancy, a criterion used to assess its conservation status, for example [6]. Depending on the quality of the input data and modelling parameters chosen, the predictions created may not forecast precisely the species’ distribution [7, 8]. Known as commission (false positives) and omission (false negatives) errors, they can inflate or reduce the potential distribution of a given taxon. Minimizing such errors is a challenge for spatial modelling [4, 7, 9], but essential considering the uses and consequences such models may have [10]. Therefore, in situ validation of the SDM outputs should be a critical step—in some cases, urgent [10, 11] since unvalidated species potential distribution maps can influence and hinder species assessments and the decision-making for species conservation.

Certifying that a species is present in a given area is not always straightforward, and different and innovative approaches have been proposed for such tasks [e.g., e-DNA [12] or satellite images [13]]. Bioacoustics is one of such techniques and has been used for a long time to record species presence/absence for amphibians, birds, and cetaceans [14]. Bats are a widespread group, using many habitats and resources, and most of the 1400 known species depend on echolocation for navigation and food acquisition [15]. The bats’ echolocation system is based on ultrasonic signals which, although above the human hearing capacity, can be easily recorded thanks to recent electronics advances [16]. Moreover, with the exception of the Phyllostomidae family, most bats have conspicuous and species-specific calls, allowing precise identification of the emitter’s identity and making the record of its presence accurate [16, 17]. Therefore, bioacoustics can be a useful technique applied to the in situ validation and refinement of SDM outputs for echolocating bats.

Here, we used an extensive in situ monitoring (>120 validation points over an area of >758,000 km2, and producing >300,000 sound files) of echolocation calls to validate the outputs of the most used SDM algorithm [MaxEnt; 3] for six neotropical bat species in a poorly-sampled part of Brazil; aiming to the evaluate SDM’s performance of using different thresholds given the validation dataset obtained with bioacoustics. Using independent acoustic data collected, we (a) evaluated the ability of four threshold-dependent theoretical evaluation metrics in predicting models’ performance, and (b) assessed the ability of three widely used thresholds to convert continuous species habitat suitability models into binary (presence/absence) maps. This methodological procedure enabled us to assess of the role of validation methods in SDM outputs and acoustic samplings as rapid validation method.

Materials and methods

Historical species records

We selected six neotropical bat species whose echolocation calls are well-known, species-specific and unequivocally identifiable: Noctilio leporinus, Promops centralis, Promops nasutus, Pteronotus gymnonotus, Pteronotus personatus, and Saccopteryx leptura [1820]. We gathered distribution records for these species from a bibliographic revision using the following online databases and search engines: the Vertebrate Zoology Database of the American Museum of Natural History (https://sci-web-002.amnh.org/db/emuwebamnh/index.php); the database of the Division of Mammals Collections of the Smithsonian National Museum of Natural History (https://collections.nmnh.si.edu/search/mammals/); the Global Biodiversity Information Facility (www.gbif.org); the SpeciesLink network (http://www.splink.org.br); Google Scholar (scholar.google.com); the Web of Science (www.webofknowledge.com); Scopus (www.scopus.com); the Periódicos CAPES (www.periodicos.capes.gov.br); and the Scientific Electronic Library Online (www.scielo.br). For that review, we searched for publications using keywords in the search engines such as: “neotropical bats”, “Pteronotus”, “Noctilio”, “Promops”, “Saccopteryx”, “Pteronotus personatus”, “P. personatus” and so on. We did not refined the searches by any specific field areas nor geographical areas. We also reviewed occurrences in studies available online such as Barquez, Ojeda [21] and Gardner [22]. We only selected bat records from peer-reviewed literature, books, or online databases supported by voucher specimens. Each record was checked for duplication of localities and, eventually, to correct location or taxonomy problems [23]. For example, following Gardner [22], we treated P. nasutus as a monotypic species (considering Promops nasutus ancilla, P. nasutus pamana, P. nasutus fosteri, or P. nasutus downsi as simply Promops nasutus). Overall, we gathered a total of 1277 single records for the six studied bat species: 590 records for Noctilio leporinus, 70 for Promops centralis, 71 for Promops nasutus, 120 for Pteronotus gymnonotus, 95 for Pteronotus personatus, and 331 for Saccopteryx leptura (Fig 1, and S1 Table).

Fig 1. Historical distribution records assembled from the literature review and the 129 acoustic sampling points used for the validation of SDM of six neotropical bat species (Noctilio leporinus, Promops centralis, Promops nasutus, Pteronutus gymnonotus, Pteronotus personatus, and Saccopteryx leptura).

Fig 1

Made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

Distribution modelling procedure

We used SDMtoolbox 2.4 for ArcGIS [24] to create an environmental heterogeneity map with all bioclimatic variables from WorldClim 2.0 [25]. To reduce the potential bias caused by autocorrelation, we then used the Spatial Rarefy Occurrence Data tool of the SDMtoolbox 2.4 package to delete records under the same environmental conditions within 25 km from each other [26]. Other studies usually use a 10km distance between points [e.g. 2729], but given the topographic and environmental heterogeneity of the studied region, we chose 25 km as a spatial filter to avoid eventual spatial bias on historical records [30]. This process ensured that all historical records gathered in our revision corresponded to a unique spatial sample, reducing occurrence data from 1277 to 899 single localities: 375 localities for Noctilio leporinus, 61 for Promops centralis, 65 for Promops nasutus, 87 for Pteronotus gymnonotus, 66 for Pteronotus personatus, and 245 for Saccopteryx leptura.

We used MaxEnt 3.4 [3] to generate potential species distribution models for the six selected species based on a set of variables at a 5 km2 resolution. Due to the shortfalls and constraints of absences and pseudo-absences in the species distribution knowledge, and consequently, in species distribution models, we chosen MaxEnt as it uses presence-only input data, can include both categorical and continuous covariables, and create a spatially explicit suitability map for the focal species. We used the 19 bioclimatic variables plus elevation available at the Worldclim data website [25], and Globcover 2009 [31] as a categorical variable for land cover. To reduce the multicollinearity among the predictor variables, we performed a preliminary model with all variables and checked the weight of each according to their contributions, using Jackknife tests. Next, using the Correlations and Summary Stats tool of the SDMtoolbox 2.4 package, we obtained the correlation and covariances matrices and removed highly correlated variables (i.e., those with the lowest value when the pairwise correlation was > 0.7) [32]. Therefore, we used different variables for each modelled species (S2 Table).

We used a logistic output to produce our models and obtain continuous suitability values for species habitat suitability, which varies from 0 (lowest suitability) to 1 (highest suitability) [3, 33]. Since default regularization values lead to overfitted models when using spatial filtering [29, 34], we calibrated the models with different regularization multiplier values (default 1.0, 2.0, 3.0, and 4.0) [29, 35]. The regularization multiplier is a ‘tunning’ parameter used to smooth the distribution prediction of the model, making it more regular and less overfitted [29, 34]. To generate overall predictive distribution models, we used 75% of the data for calibration and 25% for internal evaluation (testing data).

We used ten cross-validation replications to test all models and calculate confidence intervals, resulting in 40 models for each species, and a total of 240 continuous suitability models for all species. Since we aimed to use bioacoustical data to validate binary maps (presence-absence), we used three widely used thresholds to convert the continuous suitability values of each model into binary maps: lowest presence threshold (hereafter LPT) [28]; 10th percentile of the predicted values (hereafter P10) [28]; and maximum sum of sensitivity and specificity (hereafter maxSSS) [36, 37]. LPT is considered the least conservative threshold, whereas maxSSS the most [36, 38]. We produced a total of 720 binary models (i.e., 120 binary models for each species: 6 species x 120 binary models).

Using MaxEnt’s background and sample predictions for each model, we then evaluated each 120 species binary models for their predicting performance using four threshold-dependent theoretical evaluation metrics: (1) overall accuracy (hereafter OAcc) [39, 40]; (2) Cohen’s maximized kappa statistics (hereafter P-kappa) [41, 42]; (3) True Skill Statistics (hereafter TSS) [40, 43]; and, (4) Symmetric Extremal Dependence Index (hereafter SEDI) [44]. OAcc measures the model predicted accuracy using the rate of correct classifications (true positive + true negative) and ranges from 0 to 1; while P-kappa, TSS, and SEDI measure the model predicted accuracy taking into account the predicted accuracy of a by-chance model [40, 44]. P-kappa, TSS, and SEDI range from -1 to +1, where values ≤ 0 suggest a performance equal or worse than random, and as close the values get to +1, the better the prediction [40, 44]. For the math behind each theoretical evaluation metrics calculation, please consult Allouche, Tsoar [40], Cohen [41], Peirce [43], and Wunderlich, Lin [44].

As expected, the theoretical metrics did not converge on the same best models, therefore we selected the two best-scored binary models based on each metric (i.e., the two best-scored based on OAcc, the two best-scored based on P-kappa, and so on…) for each of the thresholds used (LPT, P10, and maxSSS). Then, we validated in the field the resulting 24 theoretical best binary models for each species, 144 binary models for the six studied species.

Acoustic monitoring and species identification

For the selection of the sampling points for field validation, we summed the MaxEnt’s given average potential distribution outputs of the six species to identify regions with the highest and lowest suitability of species occurrence, but without historical records. Considering those species’ potential distribution covered extensive areas, we focused our field validation on 129 randomly-selected sampling points along an area of 758,193 km2, in the Northeastern part of Brazil (Fig 1, and S3 Table). Since some Northeastern Brazil areas are not easily accessed, we pre-imposed the point selection near roads or paths accessible by, at least, an off-road vehicle.

Between March 2014 and January 2020, we employed passive acoustic monitoring to sample bat echolocation calls in the 129 sampling points (Fig 1, and S3 Table), using a combination of two SM2Bat+, two SM3BAT, and two SM4BAT-FS ultrasound recorders (Wildlife Acoustics Inc., Massachusetts, USA). We set the microphones at 45º to the ground, avoiding highly cluttered areas [16, 45, 46]. Since the highest frequency used by the studied species is ~60 kHz (Noctilio leporinus) [19], we configured the bat detectors with a minimum sampling rate of 384 kHz and 16 bit audio depth, enough to detect and record our focal species without distortions (e.g., aliasing). Each sampling point was acoustically monitored for at least two nights (from 2 to 15 nights), recording continuously from 30 minutes before sunset until 30 minutes after sunrise. The recordings were stored automatically in mono.wav format in the SD cards with a preset maximum duration of 1 minute if any sound above 7 kHz exceeded at least 6 dB. Since bat activity and the reception of the calls can be affected by weather and local conditions, we sampled only during nights with temperature > 15ºC, without strong winds (< 5 m/s) or rain [16, 45, 46].

We used Raven Pro 1.5 (The Cornell Lab of Ornithology 2014) for the acoustic analysis. To ease the acoustic analysis, we were divided the raw 1-minute sound files into 15-sec files and, subsequently, we visually inspected all files for the desired bat species calls after configuring the spectrograms to DFT equals 1024, 96% overlap, window length to 1 ms, using Hamming windows. We only analyzed sequences containing a minimum of three search calls with a good signal-to-noise ratio (> 15 dB) [47, 48]. We performed manual acoustic identification using qualitative (e.g., call structure and modulation) and quantitative parameters (e.g., frequency of maximum energy, maximum and minimum frequency, call duration, etc.), following previously published studies on neotropical bat acoustic identification [e.g., 1820, 4850]. We did not use feeding-buzzes (and calls immediately before and after) or social-calls for identification purposes. Although the chosen species have different natural histories, all are considered as common in the sampled region and can be easily detected and identified by their calls. All fieldwork procedures complied with the American Society of Mammalogists´ guidelines for the use of wild mammals in research and education [51] and were previously approved by the Brazilian Ministry of the Environment (SISBIO n.º 59743–1).

Field validation of the models

We evaluated the total 144 selected binary models for the six focal species against the results from the acoustic monitoring performed in the field. We used a confusion matrix to compare the accuracy of the binary maps, where the observed presence and absence cases from the acoustic monitoring were compared against the predicted presence and absence of the models. This procedure allowed us to quantify true positives (TP), true negatives (TN), false positives (FP; commission errors, type I errors), and false negatives (FN; omission errors, type II errors).

We used six metrics for the model performance evaluation: (1) Accuracy, to quantify how often the model is correct in the overall prediction [Accuracy = (TP + TN) / total cases]; (2) Precision, to quantify how often is the model correct when it predicts the occurrence of the species [Precision = TP / (TP + FP)]; (3) Sensitivity (true positive rate, or recall), to quantify the ability of the model to predict species occurrence [Sensitivity = TP / (TP + FN)]; (4) Specificity (true negative rate), which quantifies the ability of the model to predict species absence [Specificity = TN / (FP + TN)]; (5) geometric mean of sensitivity and specificity (g-mean) is a performance metric for imbalanced classifications, with high g-mean indicating a right balance between sensitivity and specificity [g-mean = √ (Sensitivity * Specificity)]. If the species presence classification performance is weak, the g-mean will be low even with an excellent species absence classification performance [52]; and (6) harmonic mean of precision and sensitivity [f-score = (2 * Precision * Sensitivity) / (Precision + Sensitivity)], gives the same importance to precision and sensitivity, i.e., high F-score indicates excellent model performance on the minority class [52, 53]. The commission error rate is inversely proportional to sensitivity (= 1—sensitivity), whereas the omission error rate is inversely proportional to specificity (= 1—specificity). The Fig 2 presents a flowchart summarizing the methodology applied for the modeling and model validation presented in this study.

Fig 2. Flowchart summarizing the methodology applied for the modeling and model validation for the six neotropical bat species presented in this study.

Fig 2

To assess the overall performance of the theoretical evaluation metrics (OAcc, P-kappa, TSS, and SEDI), we used the R package ‘lmerTest’ [54] to perform a mixed-effects linear models to evaluate the correlations between those scores and the post-validation performance metrics scores obtained using acoustic monitoring (accuracy, precision, sensitivity, specificity, g-mean, and f-score). Species were treated as random effect factor. To test differences in the prediction performances (using the post-validation performance metrics scores) between the thresholds used (LPT, maxSSS, and P10), we employed the Kruskal-Wallis test with Mann-Whitney pairwise post hoc test. The thresholds were tested with both all species together and separately.

Results

SDMs and acoustic field validation

The acoustic sampling performed in this study resulted in more than 1.5 TB of raw 1-minute sound files. Those raw sound files were divided in 15-sec files and subsequently found that more than 300,000 of those 15-sec sound files contained bat calls. As shown in the Fig 3, the calls of the six studied species are very conspicuous and easily to identify. We identified echolocation calls of Noctilio leporinus in 38 points (29,4%), of Promops centralis in 23 (17,8%), Promops nasutus in 44 (34,1%), Pteronotus gymnonotus in 53 (41,1%), Pteronotus personatus in 21 (16,3%), and Saccopteryx leptura in 24 of the 129 sampled points (18,6%) (S3 Table). Before validation, and as expected, we found that the threshold choice on the models have a great influence on the predicted species occurrence areas as the Fig 4 exemplifies for one of the models generated for species Promops nasutus.

Fig 3. Spectrogram of the echolocation calls from six neotropical bat species of this study (Noctilio leporinus, Promops centralis, Promops nasutus, Pteronotus gymnonotus, Pteronotus personatus, and Saccopteryx leptura).

Fig 3

Time scale (20 ms) in the upper right corner of the figure.

Fig 4. Maps from a model (Pnasutus_Cross_Reg1_Mod5) generated for species Promops nasutus.

Fig 4

(A) continuous suitability model; (B) LPT-thresholded binary map (predicted presence-absence); (C) P10- thresholded binary map (predicted presence-absence); (D) maxSSS-thresholded binary map (predicted presence-absence). Made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

After field validation, the performance scores varied considerably between the 144 binary models: accuracy varied from 0.16 to 0.81, precision varied between 0.09 and 0.59, sensitivity varied from 0.17 to 1, specificity from 0 to 0.86, g-mean from 0 to 0.75, and f-score from 0.12 to 0.60 (S4 Table). We registered the highest accuracy score (0.81) in a maxSSS thresholded map of Saccopteryx leptura, the highest precision score (0.59) in a maxSSS thresholded map of Noctilio leporinus, and the highest sensitivity score (= 1) in LPT thresholded maps of four species (Noctilio leporinus, Pteronotus gymnonotus, Promops nasutus, and Pteronotus personatus) (S4 Table and S1 File). The highest specificity score (0.86) was recorded in maxSSS thresholded maps of two species (Noctilio leporinus and Saccopteryx leptura), the highest g-mean score (0.75) in a P10 thresholded map of Saccopteryx leptura, and the highest f-score score (0.60) in a maxSSS thresholded maps of Pteronotus gymnonotus (S4 Table and S1 File).

Model evaluation vs. field validation

All theoretical model evaluation metrics analyzed exhibited a significant monotonic positive correlation with accuracy and specificity (Table 1). However, the evaluation metrics exhibited an overall significant negative correlation with sensitivity. We also found weak positive correlation between P-kappa and precision, and a very weak negative correlation between SEDI and precision. Although negative, the correlations between the majority of theoretical evaluation metrics and f-score were not significant, except for SEDI that displayed a significative moderate negative correlation with f-score. Only TSS and overall accuracy exhibited significant moderate or strong positive correlations with g-mean.

Table 1. Mixed-effects linear models results of the theoretical evaluation metrics and the post-validation performance metric scores for the distribution modelling of six neotropical bats based on bioacoustic field validation in northeastern Brazil.

Theoretical evaluation metric Validation performance metric Estimate Std. Dev. df t-value p-value
TSS Accuracy 0.57591 0.08081 141 7.127 < 0.001
Precision 0.71321 0.19701 106 3.62 < 0.001
Sensitivity -0.39601 0.05979 138 -6.623 < 0.001
Specificity 0.36451 0.04473 139 8.15 < 0.001
G-mean 0.50775 0.06001 139 8.46 < 0.001
F-score -0.06236 0.15176 71 -0.411 n.s.
OAcc Accuracy 0.94031 0.06659 139 14.122 < 0.001
Precision 1.22879 0.20580 129 5.971 < 0.001
Sensitivity -0.74413 0.04034 139 -18.45 < 0.001
Specificity 0.62894 0.02697 137 23.32 < 0.001
G-mean 0.72251 0.05290 139 13.659 < 0.001
F-score -0.29126 0.14502 29 -2.008 n.s.
P-kappa Accuracy 0.18566 0.04241 141 4.378 < 0.001
Precision 0.22173 0.09465 81 2.343 < 0.05
Sensitivity -0.11143 0.03156 138 -3.53 < 0.001
Specificity 0.11323 0.02429 140 4.662 < 0.001
G-mean 0.18451 0.03183 140 5.797 < 0.001
F-score 0.02633 0.07437 86 0.354 n.s.
SEDI Accuracy 0.24738 0.10510 94 2.354 < 0.05
Precision -0.006723 0.222993 73 -0.03 n.s.
Sensitivity -0.39010 0.07094 90 -5.488 < 0.001
Specificity 0.23899 0.05936 92 4.026 < 0.001
G-mean 0.22122 0.09098 93 2.431 < 0.05
F-score -0.51451 0.17516 94 -2.937 < 0.001

df, degreeds of freedom; n.s., not significative.

Thresholds vs. validation

Based on the output maps of all species together, LPT thresholded predictions exhibited significantly the lowest overall averaged accuracy, precision, specificity, and g-mean scores of the three thresholds tested (Fig 5, and S5 Table). While P10-based models exhibited a significantly higher averaged f-score than maxSSS-based, we found no significant differences in f-score between P10 and LPT, and between maxSSS and LPT-based models (Fig 5, and S5 Table). P10-based models obtained overall averaged sensitivity scores higher than maxSSS but lower than LPT-based models and specificity scores higher than LPT but lower than maxSSS-based models (Fig 5, and S5 Table). LPT-based predictions scored the highest averaged sensitivity but also presented low average specificity scores. Note that while LPT’s sensitivity scores near one, its specificity scores are also near zero (Fig 5). We found this same behaviour on LPT and maxSSS-based models when we analyzed all species separately (Fig 5, and S5 Table). One LPT prediction for N. leporinus (Fig 6) exemplifies this odd behavior, where omission errors are minimal, but commission errors are maximum (sensitivity = 1, specificity = 0, and g-mean = 0). Even the LPT prediction with the highest g-mean score (P. nasutus, Fig 7) represented a commission error rate of ~81% and an omission error rate of ~7% (sensitivity = 0.93, specificity = 0.19, and g-mean = 0.42). While maxSSS-based models scored the highest overall averaged accuracy and specificity, it also exhibited the lowest averaged sensitivity scores (S5 Table). Nevertheless, in contrast to LPT, maxSSS-based models’ sensitivity and specificity averaged scores are similar, thus presenting higher g-mean scores than LPT (Fig 5, and S5 Table). This was clear when the maxSSS-based models with the highest g-mean (Fig 8) represented an omission error of ~29% and a commission error of ~25% (sensitivity = 0.71, specificity = 0.75, and g-mean = 0.73). All thresholds significantly presented different specificity and specificity scores, but we found no significant differences between the overall accuracy, precision, and g-mean scores of maxSSS and P10 thresholded predictions (Fig 5, and S5 Table).

Fig 5. Box-plots of the accuracy, precision, sensitivity, specificity, g-mean, and f-score scores of LPT, maxSSS, and P10 thresholded maps.

Fig 5

Lower and top box boundaries represent the 25th and 75th percentile, respectively. The point inside the box represents the average, while the line represents the median. Error lines represent 1.5*interquartile range, and any asterisk outside the error lines represent outliers. Different letters indicate significant differences between groups (p < 0.05).

Fig 6. Field validation results for the LPT binary map with the highest accuracy score (= 1) for Noctilio leporinus in northeastern Brazil.

Fig 6

Omission errors are minimal but commission errors are maximum (sensitivity = 1, specificity = 0, and g-mean = 0). ‘Omission’ points represent locations where the model did not predict the species occurrence, but the species was detected during the acoustics monitoring; ‘Validated’ points represent locations where the model predict the species occurrence and the species was detected during the acoustics monitoring or locations where the model did not predict the species occurrence, and the species was detected during the acoustics monitoring; ‘commission’ points represent locations where the model predict the species occurrence but the species was not detected during the acoustics monitoring. Made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

Fig 7. Field validation results for the LPT binary map with the highest g-mean score (= 0.42) for Promops nasutus in northeastern Brazil.

Fig 7

Omission errors are low but commission errors are very high (sensitivity = 0.93, specificity = 0.19). See Fig 12 caption for the explanation on omission, validation and commission points. Made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

Fig 8. Field validation results for the maxSSS binary map with the highest g-mean score (= 0.73) for Saccopteryx leptura in northeastern Brazil.

Fig 8

Omission errors and commission errors are balanced (sensitivity = 0.71, specificity = 0.75). See Fig 12 caption for the explanation on omission, validation and commission points. Made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

Analyzing species by species, LPT-based predictions once again significantly exhibited the highest averaged sensitivity scores. However, it also showed significantly the lowest averaged accuracy, specificity, and g-mean scores for all six species studied (Fig 5, and S5 Table). In the cases of N. leporinus and S. leptura (Figs 58), LPT-based predictions also significantly exhibited the lowest averaged precision and f-score, and the lowest averaged precision scores for P. nasutus and P. personatus among the three thresholds tested (Figs 5 and 7 and S5 Table). Thresholded predictions based on maxSSS significantly exhibited the lowest averaged sensitivity scores among the three thresholds tested for N. leporinus and P. gymnonotus (Fig 5, and S5 Table). Still, maxSSS-based predictions also showed significantly the highest averaged specificity and g-mean scores for the same species. The maxSSS-based predictions also presented the highest averaged accuracy for N. leporinus and P. personatus, and the highest averaged g-mean for S. leptura among the three thresholds tested (Fig 5, and S5 Table). Saccopteryx leptura’s P10 predictions significantly exhibited the highest averaged g-mean of the three thresholds tested (Fig 5, and S5 Table). We found no differences between the performance scores based on maxSSS and P10 in all P. centralis and P. nasutus predictions (Fig 5). We also found no differences between the three thresholds in precision and f-score for P. centralis and P. personatus, f-score of P. nasutus, and precision scores of P. gymnonotus maps (Fig 5). See S1 File for the best performing binary predictions for the six species.

Discussion

This study evaluated and validated species binary distribution models using a combination of acoustic data collected in the field and simple performance metrics. We proved that bioacoustics are a very effective method for the in situ validation of SDM, here we tested it for six neotropical bat species in a large and poorly-sampled area in Brazil. Species distribution modeling is subject to interpretation and depends on many factors, such as tuning parameters and thresholds [7, 8, 29, 35, 36, 38]. After using field validation, we urge modelers to explore their modelling approaches effects on result models since predictions might be very variable even when applying optimal practices.

After modelling the distribution of six bat species, we used acoustical samplings to confirm the predicted occurrence and absences forecasted by 144 distribution potential models. For species with conspicuous vocalizations—like most bat families—the acoustical samplings we employed showed the potential to better refine SDMs in large and under-sampled regions, with relatively low sampling effort. This is quite useful in tropical areas, usually rich in bat species, but frequently understudied [55, 56]. Using bioacoustics as a validation method, we also demonstrated that a careful decision on the modelling parameters and thresholds used is pivotal since, depending on the combination, they can produce very different outputs. Our observations highlight the importance and necessity of in situ validation of SDM outputs.

Field validation of SDM is unusual [e.g., 57, 58], and very rare for bats [e.g., 10, 59, 60]. Still, we empirically demonstrated that independent field surveys are the best approach to corroborate the predictions made by modelling, especially in subsampled regions with high biodiversity like the Neotropics. We demonstrated that only using pre-existing data subsets and theoretical evaluation metrics (TSS, overall accuracy, P-kappa, and SEDI), our model evaluation may be restricted to its particularities and constraints. Simply selecting the model with the highest score may not mean that we are choosing the model that best represents reality. For example, a maxSSS-based model (Nleporinus_Cross_Reg1_Mod4_maxSSS) obtained the best TSS score for Noctilio leporinus but also displayed very high omission errors (>30%) after we validated it on the field. Although all theoretical evaluation metrics correlated positively with specificity (true negative rate) and g-mean (good balance between specificity and sensitivity), they also correlated negatively with sensitivity (true positive rate). Thus, high theoretical evaluation metric’ scores such as the widely used TSS may guarantee reliable models but could not provide the best model for our purposes and may also tend to score higher for the most overfitted models. This problem can be exacerbated when dealing with species with fewer existing data records, such as Promops centralis, and highlights how crucial in situ validation of SDMs may be.

We are aware that in situ validation of SDM is not always possible, as that will depend on the focal species, its extension of occurrence, survey methods, and the type and accessibility of the potential area modelled. In case of the echolocating bats, acoustic surveys can provide a reliable and fast validation method for distribution models. However, field validation of SDM—in smaller or focally-selected parts of the predicted distribution, or randomly selected regions—should be imperative. This is especially important in a conservation-focused scenario dealing with such high habitat changes due to anthropogenic causes. Modelling species distributions without proper in situ validation may result in inaccurate outputs, compromising the implementation of better conservation policies or species management plans, for example [5, 9, 60, 61]. This can be particularly serious in the case of models with actual low sensibility (high omission errors).

Theoretical model evaluation metrics and thresholds vs. validation

We found that all theoretical model evaluation metrics studied here correlated positively with accuracy. However, caution is necessary, since the most detected species (P. gymnonotus) was recorded in only 41% of the sampled points (meaning an unbalanced class data, i.e., in this case, absences are higher than presences). In situations like this, any random model predicting more absences than presences would be benefited by an evaluation metric that does not take into account results by chance—this is the case of accuracy [62]. Therefore, accuracy should not be used when data used to train and/or test the models is unbalanced. Here, we empirically demonstrated that sensitivity-specificity and precision-sensitivity metrics, as g-mean and f-score, are better performance measures for the SDMs evaluation than accuracy [52, 62]. For example, the Pteronotus gymnonotus’ distribution output with the highest accuracy also had the third-highest omission rate. In opposition, the model with the highest g-mean and f-score also presented low omission scores. These results were not threshold-related since they occurred in two different maxSSS models, evidencing the unbalanced nature of our SDM outputs and the problem of using accuracy to measure model performance. Considering the majority of SDMs are based on unbalanced data, instead of accuracy, the use of sensitivity-specificity and precision-sensitivity metrics should be mandatory to test the models as they are not affected by unbalanced class data sets [52, 62].

Surprisingly, we also found that all theoretical threshold-dependent evaluation metrics tested here exhibited an overall significant negative correlation with sensitivity and a significant positive correlation with specificity. This means that models with higher evaluation scores predicted better locations with actual species absence than species presence. West, Kumar [63] reported similar findings after field validation of MaxEnt’s invasive cheatgrass species models. This is probably because bioclimatic variables’ values are more homogeneous in species presence locations than in absence. Nevertheless, this ’issue’ will be less a concern if the modeler’s goal is to balance actual presences and absences, as we found positive correlations between evaluation metrics and g-mean or f-score. So, how about the ability of four threshold-dependent theoretical evaluation metrics in predicting models’ performance? In conclusion, the four theoretical model evaluation metrics studied here presented different performances about their capacity to predict the performance of models. Given the accuracy drawbacks when unbalanced data is in the game, metrics such as g-mean and f-score gain greater relevance than accuracy itself. Thus, TSS and overall accuracy (OAcc) obtained a higher ability to predict the performance of models regarding the balance between sensitivity and specificity (g-mean), however OAcc was the second-worst in terms of f-score. Still, given the strong positive correlation of most theoretical model evaluation metrics concerning specificity and negative concerning sensitivity, there is a great tendency of these metrics to better evaluate models with high omission rates. In conclusion, the choice of theoretical evaluation metrics to evaluate the models’ performance might have a great impact on its selection, and it needs to be chosen carefully towards the proposed objective.

And how about the ability of the three studied thresholds to convert continuous species habitat suitability models into presence-absence maps? We found that threshold performances varied largely. Despite having almost no omission errors, the LPT models exhibited higher commission error rates and lower accuracy, g-mean, and precision scores. Thus, at least for the six widespread neotropical bat species studied here, we were able to empirically confirm Liu, White [37]’s findings: maps based on the LPT threshold are unsuitable for species distribution modelling. In our study, LPT-thresholded maps of the two species with the most historical records (N. leporinus and S. leptura) also had a worse performance than the other four LPT-modeled scores. Although widespread and common, N. leporinus is a piscivorous bat species, strongly related to water bodies, and S. leptura is a forest-dwelling species that forage next to edges [64, 65]. Thus, these two species as less generalist than the other four we analyzed, and the use of a less conservative threshold can be detrimental in those cases. Therefore, knowledge of the species’ natural history and the use of land cover data in the models might be fundamental for best SDM practices and better output results [66]. But contrary to Liu, White [37], we also found that some maxSSS maps with high accuracy scores were highly overfitted (exhibiting low sensitivity/high omission rates), sometimes excluding historical locations for some species. Nevertheless, we also found that some maxSSS and P10 thresholded maps performed reasonably well, exhibiting balanced results between sensitivity and specificity (displaying high g-mean scores). In conclusion, we believe LPT is unsuitable for species distribution modelling, because although it reduces the omission errors of the presence-absence maps to almost zero, it does so at the expense of high commission errors. However, even showing higher omission errors than LPT, P10 thresholded models performed best in predicting actual species occurrences, while maxSSS models performed best in predicting where we did not record the species in the field. Several authors agree that threshold selection (as other parameters) has a high impact on the binary map (presence/absence) outputs and the models’ predictive capacity [e.g., 6, 8]. Several thresholds have been proposed and evaluated; however, most of those evaluations are based on theoretical evaluation metrics without independent field validation data. After using in situ validation in our study, we are cautionary about some studies still proposing to model species distribution using a single threshold.

Finally, we found that using arbitrary thresholds and theoretical evaluation metrics for modelling and evaluate the models’ performance can be a precarious approach with many possible outcomes, even if getting good evaluation scores. Validating the models using a subset of the historical occurrence points is undoubtedly faster and less laborious than using independent data collected in the field. However, one cannot guarantee if the species are still present in historical points in databases such as GBIF [67]. Hence, using independent field data is the safest way to validate the species´ presence in the modelled region and should not be overlooked.

Supporting information

S1 Table. Location of the 1277 single records of the Noctilio leporinus, Promops centralis, Promops nasutus, Pteronotus gymnonotus, Pteronotus personatus and Saccopteryx leptura, gathered after our bibliographic revision.

(XLSX)

S2 Table. Prediction variables used in this study for the distribution modelling of six neotropical bat species in northeastern Brazil.

(XLSX)

S3 Table. Presence/absence obtained in the acoustic monitoring of 129 acoustic sampling points used for the validation of SDM’s of six neotropical bat species in northeastern Brazil.

(XLSX)

S4 Table. Theoretical model evaluation scores, validation confusion matrix, and performance scores of the 144 binary distribution models of six neotropical bat species validated with field acoustics sampling in northeastern Brazil.

(XLSX)

S5 Table. Average ± standard deviation scores of accuracy, precision, sensitivity, specificity, g-mean, and f-score of the three thresholds tested for SDMs of six neotropical bat species after a field validation in northeastern Brazil using bioacoustics.

Kruskal-Wallis test results for the comparisons between the thresholds’ performance scores is also presented. The Mann-Whitney pairwise post hoc test results are presented by letters next to average ± standard deviation (different letters indicate significant differences between groups, p < 0.05).

(DOCX)

S1 File. Field validation results for the binary maps with the highest accuracy, precision, sensitivity, specificity, gmean, and f-score scores for the six studied species in northeastern Brazil.

Maps made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

(PDF)

Acknowledgments

We thank all those who helped during fieldwork. We are also grateful to ICMBio and the surveyed conservation areas’ staff for all the support. FH thanks the Raven Team Bioacoustics Research Program for the free copy of Raven software. This manuscript was based on one of the chapters of FH’s PhD Thesis at PPGBA/UFPE, and we thank the five doctoral examiners for their constructive comments during the defense.

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

FH was supported by a PhD grant from the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior–Brasil (CAPES - https://www.gov.br/capes/pt-br) – Finance Code 001 EB and RBM have produtivity fellowships conceded by the Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq - https://www.gov.br/cnpq/pt-br). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Zimmermann NE, Edwards TC Jr, Graham CH, Pearman PB, Svenning J-C. New trends in species distribution modelling. Ecography. 2010;33(6):985–9. doi: 10.1111/j.1600-0587.2010.06953.x [DOI] [Google Scholar]
  • 2.Domisch S, Friedrichs M, Hein T, Borgwardt F, Wetzig A, Jähnig SC, et al. Spatially explicit species distribution models: A missed opportunity in conservation planning? Divers Distrib. 2019;25(5):758–69. doi: 10.1111/ddi.12891 [DOI] [Google Scholar]
  • 3.Phillips SJ, Anderson RP, Schapire RE. Maximum entropy modeling of species geographic distributions. Ecological Modelling. 2006;190(3–4):231–59. doi: 10.1016/j.ecolmodel.2005.03.026 [DOI] [Google Scholar]
  • 4.Anderson RP. Harnessing the world’s biodiversity data: promise and peril in ecological niche modeling of species distributions. Annals of the New York Academy of Sciences. 2012;1260(1):66–80. doi: 10.1111/j.1749-6632.2011.06440.x [DOI] [PubMed] [Google Scholar]
  • 5.Guisan A, Tingley R, Baumgartner JB, Naujokaitis-Lewis I, Sutcliffe PR, Tulloch AIT, et al. Predicting species distributions for conservation decisions. Ecology Letters. 2013;16(12):1424–35. doi: 10.1111/ele.12189 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Fourcade Y, Besnard AG, Secondi J. Paintings predict the distribution of species, or the challenge of selecting environmental predictors and evaluation statistics. Global Ecology and Biogeography. 2018;27(2):245–56. doi: 10.1111/geb.12684 [DOI] [Google Scholar]
  • 7.Jiménez-Valverde A, Lobo JM, Hortal J. Not as good as they seem: the importance of concepts in species distribution modelling. Divers Distrib. 2008;14(6):885–90. doi: 10.1111/j.1472-4642.2008.00496.x [DOI] [Google Scholar]
  • 8.Benito BM, Cayuela L, Albuquerque FS. The impact of modelling choices in the predictive performance of richness maps derived from species-distribution models: guidelines to build better diversity models. Methods Ecol Evol. 2013;4(4):327–35. doi: 10.1111/2041-210x.12022 [DOI] [Google Scholar]
  • 9.Visconti P, Di Marco M, Álvarez-Romero J, Januchowski-Hartley S, Pressey R, Weeks R, et al. Effects of errors and gaps in spatial data sets on assessment of conservation progress. Conservation Biology. 2013;27(5):1000–10. doi: 10.1111/cobi.12095 [DOI] [PubMed] [Google Scholar]
  • 10.Greaves GJ, Mathieu R, Seddon PJ. Predictive modelling and ground validation of the spatial distribution of the New Zealand long-tailed bat (Chalinolobus tuberculatus). Biol Conserv. 2006;132(2):211–21. doi: 10.1016/j.biocon.2006.04.016 [DOI] [Google Scholar]
  • 11.Hipólito J, Hasui É, Viana BF. Solving problems involving the distribution of a species of unknown distribution via ecological niche modeling. Nat Conservacao. 2015;13(1):15–23. doi: 10.1016/j.ncon.2015.03.001 [DOI] [Google Scholar]
  • 12.Carraro L, Mächler E, Wüthrich R, Altermatt F. Environmental DNA allows upscaling spatial patterns of biodiversity in freshwater ecosystems. Nature Communications. 2020;11(1):3585. doi: 10.1038/s41467-020-17337-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Cubaynes HC, Fretwell PT, Bamford C, Gerrish L, Jackson JA. Whales from space: Four mysticete species described using new VHR satellite imagery. Marine Mammal Science. 2019;35(2):466–91. doi: 10.1111/mms.12544 [DOI] [Google Scholar]
  • 14.Laiolo P. The emerging significance of bioacoustics in animal species conservation. Biol Conserv. 2010;143(7):1635–45. doi: 10.1016/j.biocon.2010.03.025 [DOI] [Google Scholar]
  • 15.Schnitzler H-U, Moss CF, Denzinger A. From spatial orientation to food acquisition in echolocating bats. Trends Ecol Evol. 2003;18(8):386–94. doi: 10.1016/s0169-5347(03)00185-x [DOI] [Google Scholar]
  • 16.Parsons S, Szewczak J. Detecting, recording and analysing the vocalisations of bats. Ecological and Behavioral Methods for the Study of Bats [2nd Ed]: Johns Hopkins University Press; 2009. p. 91–111. [Google Scholar]
  • 17.Yoh N, Syme P, Rocha R, Meyer CFJ, López-Baucells A. Echolocation of Central Amazonian ‘whispering’ phyllostomid bats: call design and interspecific variation. Mammal Research. 2020;65(3):583–97. doi: 10.1007/s13364-020-00503-0 [DOI] [Google Scholar]
  • 18.Barataud M, Giosa S, Leblanc F, Rufray V, Disca T, Tillon L, et al. Identification et écologie acoustique des chiroptères de Guyane Française. Le Rhinolophe. 2013;19:103–45. [Google Scholar]
  • 19.Arias-Aguilar A, Hintze F, Aguiar LMS, Rufray V, Bernard E, Pereira MJR. Who’s calling? Acoustic identification of Brazilian bats. Mammal Research. 2018;63(3):231–53. doi: 10.1007/s13364-018-0367-z [DOI] [Google Scholar]
  • 20.Hintze F, Arias-Aguilar A, Dias-Silva L, Delgado-Jaramillo M, Silva CR, Jucá T, et al. Molossid unlimited: extraordinary extension of range and unusual vocalization patterns of the bat, Promops centralis. J Mammal. 2020;101(2):417–32. doi: 10.1093/jmammal/gyz167 [DOI] [Google Scholar]
  • 21.Barquez RMD, Ojeda MM, Ricardo A. Mamíferos de Argentina: sistemática y distribución: SAREM; 2006. [Google Scholar]
  • 22.Gardner AL. Mammals of South America, volume 1: marsupials, xenarthrans, shrews, and bats: University of Chicago Press; 2008. [Google Scholar]
  • 23.Peterson AT, Soberón J, Pearson RG, Anderson RP, Martínez-Meyer E, Nakamura M, et al. Ecological niches and geographic distributions (MPB-49): Princeton University Press; 2011. [Google Scholar]
  • 24.Brown JL. SDMtoolbox: a python-based GIS toolkit for landscape genetic, biogeographic and species distribution model analyses. Methods Ecol Evol. 2014;5(7):694–700. doi: 10.1111/2041-210X.12200 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Fick SE, Hijmans RJ. WorldClim 2: new 1-km spatial resolution climate surfaces for global land areas. International Journal of Climatology. 2017;37(12):4302–15. doi: 10.1002/joc.5086 [DOI] [Google Scholar]
  • 26.Coxen CL, Frey JK, Carleton SA, Collins DP. Species distribution models for a migratory bird based on citizen science and satellite tracking data. Global Ecology and Conservation. 2017;11:298–311. doi: 10.1016/j.gecco.2017.08.001 [DOI] [Google Scholar]
  • 27.Hidalgo-Mihart MG, Cantú-Salazar L, González-Romero A, López-González CA. Historical and present distribution of coyote (Canis latrans) in Mexico and Central America. Journal of Biogeography. 2004;31(12):2025–38. doi: 10.1111/j.1365-2699.2004.01163.x [DOI] [Google Scholar]
  • 28.Pearson RG, Raxworthy CJ, Nakamura M, Townsend PA. Predicting species distributions from small numbers of occurrence records: a test case using cryptic geckos in Madagascar. Journal of Biogeography. 2007;34(1):102–17. doi: 10.1111/j.1365-2699.2006.01594.x [DOI] [Google Scholar]
  • 29.Radosavljevic A, Anderson RP. Making better Maxent models of species distributions: complexity, overfitting and evaluation. Journal of Biogeography. 2014;41(4):629–43. doi: 10.1111/jbi.12227 [DOI] [Google Scholar]
  • 30.Fourcade Y, Engler JO, Rödder D, Secondi J. Mapping Species Distributions with MAXENT Using a Geographically Biased Sample of Presence Data: A Performance Assessment of Methods for Correcting Sampling Bias. PLoS ONE. 2014;9(5):e97122. doi: 10.1371/journal.pone.0097122 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Arino O, Ramos Perez JJ, Kalogirou V, Bontemps S, Defourny P, Van Bogaert E. Global land cover map for 2009 (GlobCover 2009). ESA & UCL. 2012. [Google Scholar]
  • 32.Snedecor G, Cochran W. Statistical methods. VII Edition ed. Iowa: Iowa State University Press; 1980. doi: 10.1152/ajpheart.1980.238.5.H675 [DOI] [Google Scholar]
  • 33.Phillips SJ, Dudík M, Elith J, Graham CH, Lehmann A, Leathwick J, et al. Sample selection bias and presence-only distribution models: implications for background and pseudo-absence data. Ecological Applications. 2009;19(1):181–97. doi: 10.1890/07-2153.1 [DOI] [PubMed] [Google Scholar]
  • 34.Elith J, Phillips SJ, Hastie T, Dudík M, Chee YE, Yates CJ. A statistical explanation of MaxEnt for ecologists. Divers Distrib. 2011;17(1):43–57. doi: 10.1111/j.1472-4642.2010.00725.x [DOI] [Google Scholar]
  • 35.Merow C, Smith MJ, Silander JA. A practical guide to MaxEnt for modeling species’ distributions: what it does, and why inputs and settings matter. Ecography. 2013;36(10):1058–69. doi: 10.1111/j.1600-0587.2013.07872.x [DOI] [Google Scholar]
  • 36.Liu C, Newell G, White M. On the selection of thresholds for predicting species occurrence with presence-only data. Ecology and Evolution. 2016;6(1):337–48. doi: 10.1002/ece3.1878 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Liu C, White M, Newell G. Selecting thresholds for the prediction of species occurrence with presence-only data. Journal of Biogeography. 2013;40(4):778–89. doi: 10.1111/jbi.12058 [DOI] [Google Scholar]
  • 38.Jiménez-Valverde A, Lobo JM. Threshold criteria for conversion of probability of species presence to either–or presence–absence. Acta Oecologica. 2007;31(3):361–9. doi: 10.1016/j.actao.2007.02.001 [DOI] [Google Scholar]
  • 39.Brooner W. Land-use map accuracy criteria. Photogrammetric Engineering and Remote Sensing. 1976;42(5):671–7. [Google Scholar]
  • 40.Allouche O, Tsoar A, Kadmon R. Assessing the accuracy of species distribution models: prevalence, kappa and the true skill statistic (TSS). Journal of Applied Ecology. 2006;43(6):1223–32. doi: 10.1111/j.1365-2664.2006.01214.x [DOI] [Google Scholar]
  • 41.Cohen J. A Coefficient of Agreement for Nominal Scales. Educational and Psychological Measurement. 1960;20(1):37–46. doi: 10.1177/001316446002000104 [DOI] [Google Scholar]
  • 42.Mingyang L, Yunwei J, Kumar S, Stohlgren TJ. Modeling potential habitats for alien species Dreissena polymorpha in Continental USA. Acta Ecologica Sinica. 2008;28(9):4253–8. doi: 10.1016/S1872-2032(08)60080-3 [DOI] [Google Scholar]
  • 43.Peirce CS. The numerical measure of the success of predictions. Science. 1884;ns-4(93):453–4. doi: 10.1126/science.ns-4.93.453-a [DOI] [PubMed] [Google Scholar]
  • 44.Wunderlich RF, Lin Y-P, Anthony J, Petway JR. Two alternative evaluation metrics to replace the true skill statistic in the assessment of species distribution models. Nature Conservation. 2019;35. doi: 10.3897/natureconservation.35.33918 [DOI] [Google Scholar]
  • 45.Adams AM, Jantzen MK, Hamilton RM, Fenton MB. Do you hear what I hear? Implications of detector selection for acoustic monitoring of bats. Methods Ecol Evol. 2012;3(6):992–8. doi: 10.1111/j.2041-210X.2012.00244.x [DOI] [Google Scholar]
  • 46.Ratcliffe JM, Jakobsen L. Don’t believe the mike: behavioural, directional, and environmental impacts on recorded bat echolocation call measures. Canadian Journal of Zoology. 2018:283–8. doi: 10.1139/cjz-2017-0219 [DOI] [Google Scholar]
  • 47.Lloyd A, Law B, Goldingay R. Bat activity on riparian zones and upper slopes in Australian timber production forests and the effectiveness of riparian buffers. Biol Conserv. 2006;129(2):207–20. doi: 10.1016/j.biocon.2005.10.035 [DOI] [Google Scholar]
  • 48.Jung K, Molinari J, Kalko EKV. Driving Factors for the Evolution of Species-Specific Echolocation Call Design in New World Free-Tailed Bats (Molossidae). PLoS ONE. 2014;9(1):e85279. doi: 10.1371/journal.pone.0085279 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Jung K, Kalko EK, Helversen Ov. Echolocation calls in Central American emballonurid bats: signal design and call frequency alternation. J Zool. 2007;272(2):125–37. doi: 10.1111/j.1469-7998.2006.00250.x [DOI] [Google Scholar]
  • 50.López-Baucells A, Rocha R, Bobrowiec P, Bernard E, Palmeirim J, Meyer C. Field Guide to Amazonian Bats. López-Baucells A, Rocha R, Bobrowiec P, Bernard E, Palmeirim J, Meyer C, editors. Manaus, Brazil: Editora INPA; 2016. 168 p. [Google Scholar]
  • 51.Sikes RS II, JAB, Byman D, Danielson BJ, Eggleston J, Gannon MR, et al. Guidelines of the American Society of Mammalogists for the use of wild mammals in research and education. J Mammal. 2016;97(3):663–88. Epub 2016/05/28. doi: 10.1093/jmammal/gyw078 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Branco P, Torgo L, Ribeiro RP. A Survey of Predictive Modeling on Imbalanced Domains. ACM Comput Surv. 2016;49(2):Article 31. doi: 10.1145/2907070 [DOI] [Google Scholar]
  • 53.Daskalaki S, Kopanas I, Avouris N. Evaluation of classifiers for an uneven class distribution problem. Applied Artificial Intelligence. 2006;20(5):381–417. doi: 10.1080/08839510500313653 [DOI] [Google Scholar]
  • 54.Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest Package: Tests in Linear Mixed Effects Models. 2017. 2017;82(13):26. Epub 2017-11-29. doi: 10.18637/jss.v082.i13 [DOI] [Google Scholar]
  • 55.Bernard E, Aguiar LMS, Machado RB. Discovering the Brazilian bat fauna: a task for two centuries? Mamm Rev. 2011;41(1):23–39. doi: 10.1111/j.1365-2907.2010.00164.x [DOI] [Google Scholar]
  • 56.Delgado-Jaramillo M, Aguiar LMS, Machado RB, Bernard E. Assessing the distribution of a species-rich group in a continental-sized megadiverse country: Bats in Brazil. Divers Distrib. 2020;n/a(n/a). doi: 10.1111/ddi.13043 [DOI] [Google Scholar]
  • 57.Hertzog LR, Besnard A, Jay-Robert P. Field validation shows bias-corrected pseudo-absence selection is the best method for predictive species-distribution modelling. Divers Distrib. 2014;20(12):1403–13. doi: 10.1111/ddi.12249 [DOI] [Google Scholar]
  • 58.Giné GAF, Faria D. Combining species distribution modeling and field surveys to reappraise the geographic distribution and conservation status of the threatened thin-spined porcupine (Chaetomys subspinosus). PLoS ONE. 2018;13(11):e0207914. doi: 10.1371/journal.pone.0207914 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Rebelo H, Jones G. Ground validation of presence-only modelling with rare species: a case study on barbastelles Barbastella barbastellus (Chiroptera: Vespertilionidae). Journal of Applied Ecology. 2010;47(2):410–20. doi: 10.1111/j.1365-2664.2009.01765.x [DOI] [Google Scholar]
  • 60.Razgour O, Rebelo H, Di Febbraro M, Russo D. Painting maps with bats: species distribution modelling in bat research and conservation. Hystrix, the Italian Journal of Mammalogy. 2016;27(1). doi: 10.4404/hystrix-27.1-11753 [DOI] [Google Scholar]
  • 61.Kramer-Schadt S, Niedballa J, Pilgrim JD, Schröder B, Lindenborn J, Reinfelder V, et al. The importance of correcting for sampling bias in MaxEnt species distribution models. Divers Distrib. 2013;19(11):1366–79. doi: 10.1111/ddi.12096 [DOI] [Google Scholar]
  • 62.Ferri C, Hernández-Orallo J, Modroiu R. An experimental comparison of performance measures for classification. Pattern Recognition Letters. 2009;30(1):27–38. doi: 10.1016/j.patrec.2008.08.010 [DOI] [Google Scholar]
  • 63.West AM, Kumar S, Brown CS, Stohlgren TJ, Bromberg J. Field validation of an invasive species Maxent model. Ecological Informatics. 2016;36:126–34. doi: 10.1016/j.ecoinf.2016.11.001 [DOI] [Google Scholar]
  • 64.Hood CS, Jones JK. Noctilio leporinus. Mamm Species. 1984;(216):1–7. [Google Scholar]
  • 65.Yancey FD II, Goetze JR, Jones C. Saccopteryx leptura. Mamm Species. 1998;(582):1–3. doi: 10.2307/3504379 [DOI] [Google Scholar]
  • 66.Wilson JW, Sexton JO, Todd Jobe R, Haddad NM. The relative contribution of terrain, land cover, and vegetation structure indices to species distribution models. Biol Conserv. 2013;164:170–6. doi: 10.1016/j.biocon.2013.04.021 [DOI] [Google Scholar]
  • 67.Beck J, Böller M, Erhardt A, Schwanghart W. Spatial bias in the GBIF database and its effect on modeling species’ geographic distributions. Ecological Informatics. 2014;19:10–5. doi: 10.1016/j.ecoinf.2013.11.002 [DOI] [Google Scholar]

Decision Letter 0

Daniel de Paiva Silva

29 Apr 2021

PONE-D-21-07379

Bioacoustics for in situ validation of species distribution modelling: An example with bats in Brazil

PLOS ONE

Dear Dr. Hintze,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by August 1st 2021. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Daniel de Paiva Silva, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

  1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

  1. Please ensure that you refer to Figure 9 in your text as, if accepted, production will need this reference to link the reader to the figure.

3. We note that Figures 1,9,10,11 and A-U in your submission contain map images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth). For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission:

1.              You may seek permission from the original copyright holder of Figures 1,9,10,11 and A-U to publish the content specifically under the CC BY 4.0 license. 

We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

2.              If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

The following resources for replacing copyrighted map figures may be helpful:

USGS National Map Viewer (public domain): http://viewer.nationalmap.gov/viewer/

The Gateway to Astronaut Photography of Earth (public domain): http://eol.jsc.nasa.gov/sseop/clickmap/

Maps at the CIA (public domain): https://www.cia.gov/library/publications/the-world-factbook/index.html and https://www.cia.gov/library/publications/cia-maps-publications/index.html

NASA Earth Observatory (public domain): http://earthobservatory.nasa.gov/

Landsat: http://landsat.visibleearth.nasa.gov/

USGS EROS (Earth Resources Observatory and Science (EROS) Center) (public domain): http://eros.usgs.gov/#

Natural Earth (public domain): http://www.naturalearthdata.com/

Additional Editor Comments:

Dear Hintze et al.,

After two independent reviews, I believe your manuscript may be accepted for publication after a major review when you will have the possibility to take care of the issues raised by both reviewers. The reviewers decided on a major and a minor review. Although both reviewers raised important issues, please take special care regarding the issues raised by reviewer #2, who raised important issues related to your writing and requested you to be more explicit in the relation between the study's goals and results.

Considering the pandemic situation in Brazil, I believe a three-month (August 1st 2021) period will be more than enough for you to deliver the revised version of your text. By that time, along with the revised version of your text, please do not forget to prepare a rebuttal letter where you will explain all the decisions you took regarding the issues raised by the reviewers. Do not hesitate to resubmit earlier in case you are able to. Nonetheless, in case you need more time, please let me know.

Sincerely,

Daniel Silva

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Review for PONE-D-21-07379

In this manuscript, the authors use a sound presence-only database of neotropical bats to build presence-absence maps based on species distribution modeling (SDM) techniques, specifically using MaXent. The authors test distinct combinations of parameters and validate the dataset using an impressive independent dataset, obtained with acoustic monitoring. I appreciate the introduction and believe that sampling procedures and methodological procedures are adequate for the study. The results section seems to have some results missing, and the figures could have some more editing to facilitate reading the many findings of this study. Finally, I believe that the discussion is fine in content, but could be organized to improve interpretations, with a clear take-home message for the broad public. One last point is that although the title contains bioacoustics, I missed seeing some discussion on the cost-benefit of acoustic and non-acoustic sampling to validate SDM.

This diligent study will be a good contribution to the field. Bellow, I provide some suggestions and commentaries through the manuscript to be considered before publication.

Abstract

L. 15: estimates of species diversity?

Introduction

L. 41-46: some sentences are repeated from the abstract. Consider rephrasing them.

L. 55: Is there a spatial modeling science? Maybe spatial modeling, or just modeling, would be the method to make science (ecology) more applied, where errors should be minimized.

L. 74: I think the goal is to evaluate SDM performance given the validation dataset obtained with bioacoustics and different thresholds. Having four evaluation metrics would then be a methodological procedure.

Methods

Perhaps a flowchart with the steps and setting used to build models could be provided to have a clearer picture of the distinct models.

L. 131-132: could you provide, in few words, how does the regularization multiplier parameter work?

L. 167: The validation points seem to be more concentrated toward east, not randomly distributed across the 1000 x 1000 km area. This pattern is similar to the historical records, which may likely be associated with accessibility.

L. 181: Did the acoustic monitoring took place in all months from march 2014 – January 2020? Can you provide some information about the seasonal activity of these species?

L. 182: In line 71 you mention 300,000 files. Are these files subsets within the total continuous files? If so, please, describe how subset files were selected.

Results

* You could add a paragraph summarizing the main results found in SDM with the different settings used in the models, before validation.

* What are the differences in performance found for different regularization multiplier values?

L. 236: It would be nice to see an image of the distinct calls found in such an amazing acoustic dataset.

Figures: It would be easier to see the graphs if you provide a single panel summarizing all seven figures 2-8, where each line could be a metrics (accuracy, precision, etc.) for the distinct thresholds, and boxplots for each species would be different colors.

Discussion

* The first paragraph still lacks a clear summary of the main findings and a take-home message. For instance, it remains unclear what is behind the differences among the three thresholds used in this study. What are the fundamental differences between them, and which would theoretically provide a more informative result? It would also be interesting if you could discuss which models had best performance and if post-validation performance of these specific models had high and positive correlation.

* Correlation scores in table 1 are not strong, with most <0.5. This finding should be highlighted and discussed.

* Also, it would be interesting to know if the performance metrics are similar to studies that rely upon non-acoustic methods for validation.

L. 374-375: This is not entirely true for all species/thresholds evaluated and could be acknowledged here. In this first paragraph, you could be more specific in the take-home message. For instance, you could include information on how different thresholds may be better than others and discuss the influence of sample size and unbalanced data.

Reviewer #2: This study examines the contribution of bioacoustic tools (passive acoustic monitoring) to validate the predictions of Species Distribution Models (SDM) in six tropical bat species. Based on the comparison between theoretical evaluation metrics and post-validation performance parameters obtained from field sampling, the authors highlight the need of in situ validation of SDM and argue the use of novel acoustic techniques as rapid validation methods. The study has been properly conducted, using sound methods and a large data set, which enable the authors to successfully address the proposed goals. Overall, the manuscript is clear, well-written and presents results in an effective manner. Nevertheless, there are still a series of issues that should be carefully revised before publication. First, I strongly recommend an English revision of the whole text by a native speaker or language service, if it has not been made yet. I am not an English native speaker, but I feel this is needed to significantly improve spelling, grammar, and the general flow of the text. All across the manuscript, I included suggestions and minor questions (directly on the pdf; see attachment) that aim to increase the clarity and precision of the document. My major points are listed below.

Major points

1. Statistical analysis

My main concern is related to the statistical analyses, since some of them may be fell into pseudoreplication. As shown in Table 1, Spearman correlation tests were calculated using 144 observations (except for SEDI) that came from models of the six study species (24 models per species). Thus, these observations (validation metrics) are grouped by species and they must be not considered as independent replicates. Predictions obtained by models of the same species can likely be related. As consequence, the statistical analysis applied to examine the correlation between theoretical evaluation and in situ validation should be designed taking into account the non-independence of the observations and performed again. I recommend the use of general linear mixed-effects models (GLMM), for instance, with species as random factor. Probably, the same might be applied to Kruskal-Wallis tests that the authors should carefully review in the light of this comment about potential pseureplication.

2. Results interpretation, conclusions and goals

Despite of the fact that the manuscript is generally well-written and structured, I find that the link between results and discussion is still unclear, especially for a general audience with less experience on SDM. In results, the reader can find vast details and analyses, but there are not clear explanations about the implications of these findings. In discussion, the interpretation of the results is often presented in a general manner, hindering the general understanding of the origin of such conclusions. I recommend the authors to make an effort to clarify (in results or discussion; or even better, in both) which specific result in each case enable them to draw a particular conclusion, so that the text gain in clearness and can be accessible for a broader audience. Which specific result helps us to understand that validation is key for SDM? Which one indicates that bioacoustics is “a very effective method for the in situ validation of SDM”? Which one that “we empirically demonstrated that independent field surveys are the best approach to corroborate the predictions made by modelling”?

Moreover, in my opinion, the text would also benefit from a more clear link between the goals presented in the last paragraph of the introduction and the main conclusions presented in the first paragraph of the discussion. The paragraph presenting the study goals lacks an explicit mention to the general aim of the study. The authors did not clearly refer to a key aspect of the study: the assessment of the role of validation methods in SDM and their proposal of using bioacoustic tools as rapid validation method.

3. Methodological aspects that also require clarification

- Historical species records

Authors conducted a literature review for gathering distribution records of the study species. However, some details are missing and prevent the reader from properly understand how this review was carried out. Did authors use keywords in databases as Google Scholar, WoS or Scopus? Did they refine the search by specific field areas? How many documents met the criteria and were reviewed? How many were used to determine the historical records?

- Distribution modelling procedure

This section is well-written, clear and full of details. I particularly appreciate the modelling design applied by the authors that took into account a large number of factors and criteria, and performing a diversity of models. I only have a few minor questions (see the pdf) and a major one: While historical records are taken as presences, how did the authors treat absences and pseudo-absences in the SDM models? I think this should be clarify, considering the significant knowledge shortfalls in species distribution, and the relevant effect of absence in SDM models.

- Acoustic sampling

A large number of details and information should also be added in this section to properly describe some key points (see comments on pdf). Particularly, a key assumption of the study is that the number of sampled days was large enough to get representative data of bat activity in each site. How can we be sure that sampling a minimum of two days enable the authors to determine species presence in a given location?

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: PONE-D-21-07379_reviewer.pdf

Decision Letter 1

Daniel de Paiva Silva

1 Oct 2021

PONE-D-21-07379R1Bioacoustics for in situ validation of species distribution modelling: An example with bats in BrazilPLOS ONE

Dear Dr. Hintze,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Novermber 2nd, 2021. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Daniel de Paiva Silva, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

Dear Hintze et al,

After a new review round, I must say that you are almost there. One of the original reviewers decided for the acceptance of your manuscript, whereas the other decided for a minor review. Therefore, I will grant you all a one-month period in order to proceed with the necessary changes. In case you need more time, please let me know. Still, do not hesitate to submit earlier then the estipulated deadline in case you can.

Best regards,

Daniel Silva, PhD.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I thank the authors for addressing the previous suggestions. This is an interesting contribution combining SDM and bioacoustics, and I congratulate the authors for their work. Finally, I would just suggest adjusting some minor concerns, detailed below.

Figures 5-11: I think that you should reconsider joining all boxplots in a single figure. I suggest that you use accuracy, precision, sensitivity, specificity, g-mean and f-score as lines, and for columns, use LPT, maxSSS, and p10 for repeated for each species. Such as:

Species 1 / Species 2 / Species 3 …

LPT. maxSSS P10 / LPT. maxSSS P10/ LPT. maxSSS P10

Acc

Pre

Sen

Spe

Gme

Fsc

Species 6 / Species 7 / Species 8…

LPT. maxSSS P10 / LPT. maxSSS P10/ LPT. maxSSS P10

Acc

Pre

Sen

Spe

Gme

Fsc

Let the first row with 4 species and the second row with 3.

424: “we proved” is not exactly what you show here, since the performance of the validation is subject to interpretation and depends on many factors, as tuning parameters and thresholds. I believe that the message should be to explore variations in modeling approaches using SDM, since predictions are largely variable even for approaches that are considered optimum.

Reviewer #3: I greatly enjoyed reading this paper, and believe it constitutes an important contribution to the literature on species distribution modeling (SDM), especially in Neotropical region, particularly as it provides important insights on the validation process of SDM.

The article is interesting and overall reads well; however, I have minor comments. The validation approach (based on acoustics) is especially important for non-phyllostomid bats since bats from the Phyllostomidae family have considerable overlap in their call parameters (Yoh et al. 2020). In that sense, I suggest clarifying the main target bat species, which acoustics as validation method, is proper.

Is there any season effect on acoustic surveys explaining absences and potentially affect validation?

Lines 80 – 82. It is true for several families, but it is problematic for Phyllostomidae (Yoh et al. 2020). I suggest clarifying this limitation

Lines 428 -431… It is not true for all bats

Yoh et al, (2020) Echolocation of Central Amazonian ‘whispering’ phyllostomid bats: call design and interspecific variation. Mammal Research

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 2

Daniel de Paiva Silva

7 Oct 2021

Bioacoustics for in situ validation of species distribution modelling: An example with bats in Brazil

PONE-D-21-07379R2

Dear Dr. Hintze,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Daniel de Paiva Silva, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Dear Hintze et al.

Congratulations! I am pleased to inform you that your manuscript was formally accepted for publication in PLoS One.

Best regards,

Reviewers' comments:

Acceptance letter

Daniel de Paiva Silva

11 Oct 2021

PONE-D-21-07379R2

Bioacoustics for in situ validation of species distribution modelling: An example with bats in Brazil

Dear Dr. Hintze:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Daniel de Paiva Silva

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Location of the 1277 single records of the Noctilio leporinus, Promops centralis, Promops nasutus, Pteronotus gymnonotus, Pteronotus personatus and Saccopteryx leptura, gathered after our bibliographic revision.

    (XLSX)

    S2 Table. Prediction variables used in this study for the distribution modelling of six neotropical bat species in northeastern Brazil.

    (XLSX)

    S3 Table. Presence/absence obtained in the acoustic monitoring of 129 acoustic sampling points used for the validation of SDM’s of six neotropical bat species in northeastern Brazil.

    (XLSX)

    S4 Table. Theoretical model evaluation scores, validation confusion matrix, and performance scores of the 144 binary distribution models of six neotropical bat species validated with field acoustics sampling in northeastern Brazil.

    (XLSX)

    S5 Table. Average ± standard deviation scores of accuracy, precision, sensitivity, specificity, g-mean, and f-score of the three thresholds tested for SDMs of six neotropical bat species after a field validation in northeastern Brazil using bioacoustics.

    Kruskal-Wallis test results for the comparisons between the thresholds’ performance scores is also presented. The Mann-Whitney pairwise post hoc test results are presented by letters next to average ± standard deviation (different letters indicate significant differences between groups, p < 0.05).

    (DOCX)

    S1 File. Field validation results for the binary maps with the highest accuracy, precision, sensitivity, specificity, gmean, and f-score scores for the six studied species in northeastern Brazil.

    Maps made with Natural Earth. Free vector and raster map data @ naturalearthdata.com.

    (PDF)

    Attachment

    Submitted filename: PONE-D-21-07379_reviewer.pdf

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Attachment

    Submitted filename: Response to Reviewers_Rev2.pdf

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES