Skip to main content
PLOS One logoLink to PLOS One
. 2023 Jul 25;18(7):e0284076. doi: 10.1371/journal.pone.0284076

COWAVE: A labelled COVID-19 wave dataset for building predictive models

Melpakkam Pradeep 1, Karthik Raman 2,3,4,*
Editor: Hilal Tayara5
PMCID: PMC10368260  PMID: 37490468

Abstract

The ongoing COVID-19 pandemic has posed a significant global challenge to healthcare systems. Every country has seen multiple waves of this disease, placing a considerable strain on healthcare resources. Across the world, the pandemic has motivated diligent data collection, with an enormous amount of data being available in the public domain. In this manuscript, we collate COVID-19 case data from around the world (available on the World Health Organization (WHO) website), and provide various definitions for waves. Using these definitions to define labels, we create a labelled dataset, which can be used while building supervised learning classifiers. We also use a simple eXtreme Gradient Boosting (XGBoost) model to provide a minimum standard for future classifiers trained on this dataset and demonstrate the utility of our dataset for the prediction of (future) waves. This dataset will be a valuable resource for epidemiologists and others interested in the early prediction of future waves. The datasets are available from https://github.com/RamanLab/COWAVE/.

Introduction

The COVID-19 pandemic has been a significant threat to global public health, and has continued to spread throughout the world, wreaking havoc on the healthcare infrastructure [15]. Despite the widespread availability of numerous vaccines and treatment methods [68], new outbreaks have continued to occur. With the absence of a permanent cure, predicting the onset of waves is a very important problem, since it could allow countries to prepare their health infrastructure and policies to mitigate the disastrous effects of a COVID-19 wave.

Multiple approaches have been attempted when trying to predict COVID-19 waves. The pandemic has been notoriously difficult to predict accurately [9]. Yet, the most common approach has been modelling the disease using a susceptible-infected- removed (SIR) model or its variants [10]. Various methods of computing the parameters involved in the differential equations of populations, lead to different predictions of waves [1113]. Some models have used Bayesian Learning to estimate these parameters [14, 15]. The computation of the reproduction number (R0) forms a key challenge in nearly all of these methods. However, for all these models, the evaluation typically requires labelled data. Machine Learning classifiers can also be used to predict waves, without having to estimate the various parameters of the SIR model.

In this manuscript, we provide a cleaned labelled dataset, ‘COWAVE’. Our dataset contains labels on whether each day was part of a “wave” or not, along with multiple helpful features extracted from the data. We also provide a baseline supervised classifier that can be helpful when comparing the performance of any classifier trained on COWAVE. We believethat scientists would find COWAVEhelpful, for building models that learn the dynamics of various kinds of COVID-19 outbreaks, and use it to predict new waves of the pandemic, quickly and reliably. The rest of this manuscript is organised as follows: we first present the data source and methods for smoothing the data and defining waves for labelling the dataset. We then present the datasets generated in this study. We then go on to discuss example classifiers and their predictions, along with possibilities for feature generation and selection.

Data and methods

In this section, we describe the source of our data, various smoothing algorithms and our approach to defining waves. All codes used in this study, and the datasets generated, are available from GitHub (https://github.com/RamanLab/COWAVE/). The codes were all written in Python 3, and the analyses were performed on the cloud, using Google Colaboratory (https://colab.research.google.com/).

Source and cleaning

All data were obtained from the World Health Organization (WHO) website (https://covid19.who.int/WHO-COVID-19-global-data.csv). The columns of the dataset provide the Date Reported, the Country Name and Code, and the WHO Region, along with the new and cumulative cases and death counts. No preprocessing was performed. Two things must be noted. The first is that the new case count suddenly becomes 0 or very low for certain dates. This is likely due to no testing being done on those days/improper testing. A moving average can be used to interpolate these points, but we do not do so since smoothing will be done later.

The second thing to note is that the Country Code for Namibia is NA. This could cause certain libraries to interpret these strings as NaN values during the exploratory analysis stage.

Smoothing algorithms

Locally weighted scatter-plot smoothing

Locally Weighted Scatter-plot Smoothing (LOWESS) is a common method used for the smoothing of scatter-plot data. It is a method equivalent to the Savitzky-Golay filter and was rediscovered by Cleveland [16]. Based on a smoothing parameter α, the dataset is divided into subsets. A low-degree polynomial (linear, quadratic usually) is fitted locally to each subset. Points closer to the point whose response is estimated are given higher weights.

The only problem with LOWESS is that it does not provide the polynomials used to approximate the dataset. In a way, it is a black box. However, this is not a problem here since we are only interested in smoothing the data.

Simple exponential smoothing

The simplest of the exponential smoothing methods is called Simple Exponential Smoothing (SES). This method is suitable for forecasting data with no clear trend or seasonal pattern [17]. This also makes it an excellent method for smoothing data. Unlike a moving average where the weights are fixed, when smoothing the data by forecasting the next data point using the past data points, SES uses weights that exponentially decrease, as data points come from further in the past.

y^T+1|T=αyT+α(1-α)yT-1+α(1-α)2yT-2+ (1)

where y^T+1|T is the forecast for t = T+1 based on the data till t = T, yx is the data point at t = x and α is a smoothing parameter (0 ≤ α ≤ 1).

The wave definitions for labelling

The literature does not offer a universal definition of a “wave”. While waves are characterized by rising and falling parts, which when put together form the wave, this definition is very vague for obvious reasons.

For example, in Fig 1, visually, we can find one large wave. Simply looking for rising and falling parts leads us to multiple “waves” since the smaller fluctuations (outbreaks) that may or may not be significant also are taken into account.

Fig 1. LOWESS smoothed cases data for Italy from 3-Jan-2020 to 20-Oct-2021.

Fig 1

The notion of a wave in the context of creating a dataset for a classifier is even more ill-defined [18, 19]. Not only must the definition be reliable, but it also must only capture what can visually be observed as a “wave”. To this end, we use a number of smoothing algorithms, which we have described in the previous section.

Definition 1: Peaks and troughs (Hale)

Hale et al. [20] defined waves as regions between two peaks or troughs, for a time series, after LOWESS smoothing, provided the peaks/troughs are at least one month apart. COVID waves, however, frequently have smaller “outbreaks”, that aren’t quite waves but get captured as waves in this definition. These characteristics can be seen in Fig 2a and 2b, S1 and S2 Figs. Here, different colours indicate different waves.

Fig 2.

Fig 2

(Definition 1) LOWESS smoothed cases data for (a) Italy, (b) India. The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). Here, regions of different colours indicate different unique waves.

In Fig 2b, visually we find two “waves”, however, this definition gives us four waves! A similar problem is found with S2 Fig, where we expect two “waves” (or three, if you count the trailing end) but find six waves. Clearly, any smaller outbreaks cause this definition to break down.

From the point of view of building a dataset for a classifier, this definition presents an even bigger problem. Since the time series can always be partitioned into sections of data between minima, there is not a single stretch of data that we can say is not part of a wave!

This can partly be avoided by imposing a minimum number of cases that a day must cross to be considered a part of a wave. For example, if we considered 200 cases as the threshold, we could eliminate the first cyan and purple portions in S1 Fig as a wave. However, even with a high threshold of 20 (given the maximum is 80, this is quite high a threshold), our problems would still persist, as in the case of S2 Fig, our definition presents five waves (the purple outbreak is entirely eliminated, along with some sections of cyan, yellow, black, etc.).

Definition 2: Rapid doubling

While the previous definition is not wrong, it does not take into account the rate at which cases rise during a certain period. Waves are mainly characterized by a rapid rise until reaching the maximum, so taking this into account could lead to much better results. Outbreaks rarely have a rapid rise. While they would be included as a “Wave” in the previous definition (due to this stretch being between two minima), in this definition, they would not. This solves the no “Non-Wave” problem to a certain degree.

Here, we find the doubling rate and normalize it:

Doublingrate=1ln2·ln|xtxt-7| (2)

where x is the data point (deaths, cases, hospitalizations, etc.).

We then set a threshold for the doubling rate and label it as “Wave” if the doubling rate at that time step is greater than the threshold value. To generate Fig 3a–3c we find the doubling rate. Since we use the raw data for this, there are multiple fluctuations in the doubling rate graph. So, we smooth the doubling rate using LOWESS (1/14).

Fig 3.

Fig 3

Cases data (a) without smoothing for Italy, (b) with smoothing for Italy, (c) with smoothing for India. (Definition 2) The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the green curve smoothed data; and the red curve indicates whether the day is part of a wave. (a) The orange curve indicates the raw doubling rate, visually, two separate waves are observed, however, the red curve indicates more than ten waves. (b) The orange curve indicates the smoothed doubling rate, visually, two separate waves are observed, however, the red curve indicates ten waves. (c) The orange curve indicates the smoothed doubling rate, visually, two separate waves are observed, however, the red curve indicates four waves.

All data is then normalized by dividing by the maximum of that quantity, to help fit all quantities in the same graph. The threshold is set to dtdt,max0.02, where dt is the doubling rate at that time instant and dt,max is the maximum doubling rate for the whole time series. For days where the normalized doubling rate is greater than the threshold, we give the label “wave” (1), and for the rest, we give the label “non-wave” (0). From the figures that follow, we can observe some problems with this definition.

The first problem is that the number of waves labelled is not what is visually apparent. For example, in S3 Fig, we can visually see two waves, but the labelling definition gives us 15 waves! Secondly, near the maxima, the doubling rate dramatically decreases. This means our definition would never capture parts of the wave near the peak. Even with this definition, there are very few stretches of “non-wave”s. Visually, we can see that the wave and non-wave data points must not be very unbalanced. So, this is another problem. Also, outbreaks tend to have higher doubling rates than “waves” since outbreaks occur over shorter time periods. This definition would definitely label outbreaks as “waves”.

Definition 3

We can observe that characterizing waves by doubling rate or minimas is not very helpful. What characterizes a wave, visually, is a combination of how much deviation there is from the level of the data, and over how many days it occurs. Waves are typically over several weeks, while outbreaks are over several days, regardless of the cases/deaths/hospitalizations in their respective outbreaks and waves. So by smoothing the time series, we can eliminate most outbreaks, allowing us to focus only on the waves.

In this definition, we think of waves as a deviation from the mean/level of the data. So, we subtract the mean of the complete time series of a measure for each country from each data point of the measure. Now, outbreaks are also deviations from the mean. To eliminate them, we perform two smoothing operations on this data, first LOWESS Smoothing (1/14) and then, Simple Exponential Smoothing. We can then scale this data appropriately (dividing by max of data or by standard deviation) to fit it in graphs when plotting.

After this, we label all transformed data points <0 as “waves” (1) and the rest as “non-waves” (0). So, the labelling definition is,

f(xt)={1ifxt-xt,mean>00else (3)

where xt is the number of cases/deaths/ hospitalizations on day t, xt,mean is the mean of the number of cases/deaths/ hospitalizations for that country. Since the initial parts of waves may not be captured, we apply a correction factor. We divide the labelled “wavelength” by a factor k and retroactively label all points within length wavelengthk from the start of the labelled wave as “wave” (1). The current value is k = 6 but different values can be experimented with for better results.

In this definition, what we see is in near-perfect agreement with the labelled waves in Fig 4a and 4b, S5 and S6 Figs! One drawback is the fact that for certain waves, it appears that the wave labelling starts too late. However, for waves with initial rapid rises (most likely to overwhelm medical infrastructure), the labelling is extremely close to the start of the wave.

Fig 4.

Fig 4

(Definition 3) LOWESS smoothed cases data for (a) Italy, (b) India. The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the orange curve, smoothed data, and the green curve indicates whether the day is part of a wave. Visually, two separate waves are observed, and the green curve also indicates two waves for both (a) and (b).

We also obtained significant “non-wave” portions, as expected. This definition could be further improved by changing k or subtracting x ⋅ mean, where 0 < x < 1 instead of just the mean, but we shall use this definition as our labelling rule for the datasets.

Results

We now present the datasets that we generate in our study, as well as the results of a few classification experiments, to provide baselines for further analysis and experimentation.

The datasets

We present curated datasets in two forms, for the convenience of analysis.

Dataset v1

This dataset consists of four columns: Date_reported, Country_code, New_cases and Wave. The Wave column is generated as described in Definition 3, with 1 standing for the day being part of a wave, and 0 meaning the day is not part of a wave.

Dataset v2

This dataset consists of stretches of waves and non-waves. Each row contains the start date of the wave/non-wave (Date), the country code (Country_code), whether it is a wave or non-wave (Wave), and the case data for that stretch of wave/non-wave (as a list) (Cases).

Training and test datasets

For the classifiers built in the next section, the training dataset consists of the WHO cases data for all countries between country codes AF (Afghanistan) and QA (Qatar) from 3/1/2020 to 27/5/2022. The test dataset consists of the WHO cases data for all countries between country codes RK (Republic of Korea) and ZW (Zimbabwe) from 3/1/2020 to 27/5/2022. All metrics are reported for the full test and training datasets. They may vary for different subsets of these datasets. We assume that the distributions of cases do not depend on the country code assigned by WHO.

A naïve baseline classifier

The baseline classifier is a simple classifier that we can use to compare all other classifiers. For this, we choose an SVM classifier [21] with a Radial Basis Function (RBF) kernel.

We first see how fast the untuned model’s metrics vary as more data is used for every iteration. Instead of sending each day’s data, we send a data vector with the current day’s cases, along with the previous six days as well. In the next run, we send in a 14-day vector instead of the 7-day vector. In the subsequent run, we send in a 21-day vector. The results of the three runs are tabulated in Table 1.

Table 1. Performance of different data vectors.

Shown are the accuracy, recall and precision for the untuned baseline classifier (SVM, RBF kernel) with the three data vectors described in this section.

Dataset n-day vector Accuracy Recall Precision
Train 7-day vector 0.686 0.154 0.722
14-day vector 0.698 0.169 0.807
21-day vector 0.705 0.184 0.839
Test 7-day vector 0.647 0.146 0.614
14-day vector 0.651 0.145 0.641
21-day vector 0.654 0.156 0.649

XGBoost provides much better classification performance

For an improved classifier, we look for an algorithm that runs quite fast and can classify tabulated data with good accuracy (>60%) and high recall. It also should not require a very large amount of data. While these requirements have not been perfectly quantified here, tree ensemble methods seem to fit all these requirements. For this, we use the XGBoost algorithm [27].

We first see how fast the untuned model’s metrics vary as more data is used for every iteration. Instead of sending each day’s data, we send a data vector with the current day’s cases, along with the previous six days as well. In the next run, we send in a 14-day vector instead of the 7-day vector. In the subsequent run, we send in a 21-day vector. The results of the three runs are tabulated in Table 2. It is evident that the XGBoost classifier gives significantly better results than the Baseline Classifier.

Table 2. Performance of different data vectors.

Shown are the accuracy, recall and precision for the untuned improved classifier (XGBoost) with the three data vectors described in this section.

Dataset n-day vector Accuracy Recall Precision
Train 7-day vector 0.735 0.538 0.665
14-day vector 0.735 0.539 0.666
21-day vector 0.737 0.542 0.669
Test 7-day vector 0.725 0.444 0.673
14-day vector 0.726 0.445 0.676
21-day vector 0.727 0.444 0.680

Feature generation and selection

Feature generation

We can clearly see that while the metrics show an improvement, as a whole, as the length of the data vector is increased, the improvements are very marginal. It is likely that we will need to generate more features, not simply use the previous days’ values as additional features if improvements are to be made to the performance. Some obvious features are the differences in the lags. We include the differences over the past seven days. Improvements have been observed to be marginal beyond this. Taking the maximum, minimum, mean and variance of the 21-day vector are also features that can be readily included.

We then split the time series into its trend, seasonal and residual components and use these as three additional features. Visualizing the time series in different ways by transforming it could also lead to better model performance in our classification task. We use the square, square root, logarithm and Box-Cox transformations here. Other transformations can also be used. These transformations need not be limited to the current-day data but can also be used for the lags and the differences.

We also include non-linear transformations of the current day’s data as features, particularly using

f(t21)=11+e-t21σ(ti) (4)

where t21 is the current day’s data and σ(ti) is the standard deviation of the 21-day vector. If the standard deviation is 0, we take the exponential term as 0.

The second non-linear transformation used is

f(t21)=12πσe-(t21-μ)22σ2 (5)

where t21 is the current day’s data, μ is the mean of the 21-day vector and σ is the standard deviation of the 21-day vector.

We also find the range and covariance of each 21-day vector since these quantities give us a measure of the spread of the cases in the 21 days. The median and entropy for the 21-day vectors are also found, where the entropy (E) is given by,

E=-i=121xiln(xi),xi=timax(t1,t2,,t21) (6)

Feature selection

The generated features were used to evaluate the importance of the generated features using an XGBoost Classifier. Based on this, the top 13 features were selected. The feature names are listed here: T21, D7, MIN, MAX, Range, Sq, Median, Mean, Variance, PDF, Trend, Seasonal and Residual. This is done to reduce the training time for the models. From now, this set of features is referenced as “Top 13”. Other feature selection methods may also be used [2226].

New features enable improved classification

Untuned XGBoost classifier

With these features, we train an XGBoost model on the dataset. With the default hyperparameters (as in the xgboost library), we achieve scores as shown in Table 3.

Table 3. Performance of different feature sets.

Shown are the accuracy, recall and precision for the untuned XGBoost classifier [27] with two subsets of the generated features—all features as well as “Top 13” features described earlier.

Dataset Features Used Accuracy Recall Precision
Train All 0.859 0.798 0.810
“Top 13” 0.855 0.796 0.801
Test All 0.830 0.690 0.804
“Top 13” 0.831 0.702 0.800

For tuning, we used Bayesian Optimization as implemented in the bayesian-optimization library [28]. Its speed, along with its ability to search the hyperparameter space, rather than discrete values only, were the reasons this method was chosen over GridSearch. We used three-fold cross-validation. The list of tuned hyperparameters, along with the tuning ranges, is given in Table 4.

Table 4. Search space for hyperparameter tuning.
Hyperparameter Value Range
learning_rate (0.0005, 1)
max_depth (1, 10)
min_child_weight (1, 10)
gamma (0, 3)
colsample_bytree (0.001, 1)
num_boost_round (100, 500)
reg_lambda (0.01, 10)
scale_pos_weight (1, 10)
subsample (0.001, 1)

Multiple runs were tested, and the best set of hyperparameters is shown here. However, due to the random nature of Bayesian Optimization, there is a chance of a better set of hyperparameters existing in a different tuning space or even in the same tuning space.

We present the hyperparameters and results for two classifiers, the first for the maximum accuracy obtained, and the second, biased towards a high recall, with acceptable accuracy (For example, models with 0.999 recall but with <0.5 accuracy would be rejected).

High accuracy classifier

Initially, the hyperparameter search was performed for the selected features. However, a higher accuracy was obtained for the tuned model when including all generated features. We present the chosen hyperparameters in Table 5, along with the results in Table 6, for both sets of features here. The classifier with the selected features runs twice as fast when compared to the classifier with all generated features.

Table 5. Selected hyperparameters for the high accuracy classifier.

The first column contains the best hyperparameters for the “Top 13” features, as described in the previous section. The second column contains the best hyperparameters for all features generated.

Hyperparameter Selected Values
“Top 13” All features
learning_rate 0.25 0.709
max_depth 4 3
min_child_weight 9.5 2.791
gamma 2.1 2.106
colsample_bytree 0.9 0.619
num_boost_round 450 336
reg_lambda 5.36 4.515
scale_pos_weight 1.0 1.437
subsample 1.0 0.504
Table 6. Performance of different feature sets.

Shown are the accuracy, recall and precision for the tuned XGBoost classifier with two subsets of the generated features—all features as well as “Top 13” Features described previously. The hyperparameters are as described in Table 5.

Dataset Feature Set used Accuracy Recall Precision
Train “Top 13” 0.867 0.820 0.814
All 0.885 0.878 0.817
Test “Top 13” 0.834 0.716 0.799
All 0.850 0.769 0.802

High recall classifier

When predicting waves, the cost of missing waves is much more than wrongly predicting waves. Therefore, we tune our models for higher recall. Since strictly maximum recall is not our aim, we use the selected features (since it is faster). Based on Bayesian Optimization, the hyperparameters were chosen as given in Table 7. With this tuned model, the results for the full test and train sets are given in Table 8. With a faster runtime, and better accuracy and recall, the XGBoost model is much better when compared to the tuned baseline.

Table 7. Selected hyperparameters for the high recall classifier.

The hyperparameters are selected based on the performance of the “Top 13” features described earlier.

Hyperparameter Value
learning_rate 0.025
max_depth 9
min_child_weight 6.441
gamma 2.444
colsample_bytree 0.873
num_boost_round 307
reg_lambda 5.897
scale_pos_weight 8.198
subsample 0.174
Table 8. Performance of different feature sets.

Shown are the accuracy, recall and precision for the tuned XGBoost classifier with the “Top 13” Features described earlier. The hyperparameters are as described in Table 7.

Dataset Accuracy Recall Precision
Train 0.777 0.977 0.622
Test 0.794 0.933 0.646

Conclusion

In this study, we present a new definition for demarcating “waves” in COVID-19 outbreaks. Our definition is simple and also agrees with typical visual interpretations of COVID-19 waves. We used this definition to label the daily case data of COVID-19 as “Wave” and “Non-Wave”, to allow the use of supervised learning classifiers to predict waves. To illustrate possibilities, we build a naïve SVM-based classifier and an improved classifier based on XGBoost. We then go on to generate new features and select more informative features, and build classifiers that are specifically tuned for higher accuracy, or more importantly, higher recall. Overall, we believe that this dataset will motivate more people to work on this exciting problem of COVID-19 wave detection, and enable the creation of more high-performing classifiers to accurately, rapidly and reliably predict future COVID waves.

Supporting information

S1 Fig. LOWESS smoothed cases data for the Netherlands.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). Here, regions of different colours indicate different unique waves.

(TIF)

S2 Fig. LOWESS smoothed cases data for Egypt.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). Here, regions of different colours indicate different unique waves.

(TIF)

S3 Fig. LOWESS smoothed cases data for the Netherlands.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the green curve, smoothed data; the orange curve, the LOWESS smoothed doubling rate, and the red curve indicates whether the day is part of a wave.

(TIF)

S4 Fig. LOWESS smoothed cases data for Egypt.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the green curve, smoothed data; the orange curve, the LOWESS smoothed doubling rate, and the red curve indicates whether the day is part of a wave.

(TIF)

S5 Fig. LOWESS smoothed cases data for the Netherlands.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the orange curve, smoothed data, and the green curve indicates whether the day is part of a wave.

(TIF)

S6 Fig. LOWESS smoothed cases data for Egypt.

The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the orange curve, smoothed data, and the green curve indicates whether the day is part of a wave.

(TIF)

Acknowledgments

We thank Dr. G. Ramadurai and other members of the IIT Madras RBCDSAI “COVID War Room Team” for several useful discussions.

Data Availability

The data underlying the results presented in the study are available from the GitHub repository (https://github.com/RamanLab/COWAVE).

Funding Statement

Funding support from IBSE and RBCDSAI is gratefully acknowledged.

References

  • 1. Nkengasong JN, Mankoula W. Looming threat of COVID-19 infection in Africa: act collectively, and fast. Lancet (London, England). 2020;395(10227):841–842. doi: 10.1016/S0140-6736(20)30464-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Pereira M, Oliveira AM. Poverty and food insecurity may increase as the threat of COVID-19 spreads. Public Health Nutrition. 2020;23(17):3236–3240. doi: 10.1017/S1368980020003493 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Petrino R, Riesgo LGC, Yilmaz B. Burnout in emergency medicine professionals after 2 years of the COVID-19 pandemic: a threat to the healthcare system? European Journal of Emergency Medicine. 2022;29(4):279–284. doi: 10.1097/MEJ.0000000000000952 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Brown J, Guru S, Williams K, Florentino R, Miner J, Cagir B. Rural Healthcare Center Preparation and Readiness Response to Threat of COVID-19. Journal of the American College of Surgeons. 2020;230(6):1105–1110. doi: 10.1016/j.jamcollsurg.2020.04.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Kaye AD, Okeagu CN, Pham AD, Silva RA, Hurley JJ, Arron BL, et al. Economic impact of COVID-19 pandemic on healthcare facilities and systems: International perspectives. Best Practice & Research Clinical Anaesthesiology. 2021;35(3):293–306. doi: 10.1016/j.bpa.2020.11.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Rizk JG, Barr CE, Rizk Y, Lewin JC. The next frontier in vaccine safety and VAERS: Lessons from COVID-19 and ten recommendations for action. Vaccine. 2021;39(41):6017–6018. doi: 10.1016/j.vaccine.2021.08.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. El Kantar S, Nehmeh B, Saad P, Mitri G, Estephan C, Mroueh M, et al. Derivatization and combination therapy of current COVID-19 therapeutic agents: a review of mechanistic pathways, adverse effects, and binding sites. Drug Discovery Today. 2020;25(10):1822–1838. doi: 10.1016/j.drudis.2020.08.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Issa H, Eid AH, Berry B, Takhviji V, Khosravi A, Mantash S, et al. Combination of Angiotensin (1-7) Agonists and Convalescent Plasma as a New Strategy to Overcome Angiotensin Converting Enzyme 2 (ACE2) Inhibition for the Treatment of COVID-19. Frontiers in Medicine. 2021;8:620990. doi: 10.3389/fmed.2021.620990 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Roda WC, Varughese MB, Han D, Li MY. Why is it difficult to accurately predict the COVID-19 epidemic? Infectious Disease Modelling. 2020;5:271–281. doi: 10.1016/j.idm.2020.03.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Agrawal M, Kanitkar M, Phillip D, Hajra T, Singh A, Singh A, et al. SUTRA: A Novel Approach to Modelling Pandemics with Applications to COVID-19; 2022. Available from: http://arxiv.org/abs/2101.09158. [Google Scholar]
  • 11. Alser M, Kim JS, Almadhoun Alserr N, Tell SW, Mutlu O. COVIDHunter: COVID-19 Pandemic Wave Prediction and Mitigation via Seasonality Aware Modeling. Frontiers in Public Health. 2022;10. doi: 10.3389/fpubh.2022.877621 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Kaxiras E, Neofotistos G. Multiple Epidemic Wave Model of the COVID-19 Pandemic: Modeling Study. Journal of Medical Internet Research. 2020;22(7):e20912. doi: 10.2196/20912 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Nesteruk I. Waves of COVID-19 pandemic. Detection and SIR simulations; 2020. Available from: https://www.medrxiv.org/content/10.1101/2020.08.03.20167098v1.
  • 14. Braca P, Gaglione D, Marano S, Millefiori LM, Willett P, Pattipati KR. Quickest Detection of COVID-19 Pandemic Onset. IEEE Signal Processing Letters. 2021;28:683–687. doi: 10.1109/LSP.2021.3068072 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Soldi G, Forti N, Gaglione D, Braca P, Millefiori LM, Marano S, et al. Quickest Detection and Forecast of Pandemic Outbreaks: Analysis of COVID-19 Waves. IEEE Communications Magazine. 2021;59(9):16–22. doi: 10.1109/MCOM.101.2001252 [DOI] [Google Scholar]
  • 16. Cleveland WS. LOWESS: A Program for Smoothing Scatterplots by Robust Locally Weighted Regression. The American Statistician. 1981;35(1):54. doi: 10.2307/2683591 [DOI] [Google Scholar]
  • 17. Hyndman RJ, Athanasopoulos G. Forecasting: Principles and Practice. OTexts. 2018. [Google Scholar]
  • 18. Arroyo-Marioli F, Bullano F, Kucinskas S, Rondón-Moreno C. Tracking R of COVID-19: A new real-time estimation using the Kalman filter. PLOS ONE. 2021;16(1):e0244474. doi: 10.1371/journal.pone.0244474 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Zhang SX, Arroyo Marioli F, Gao R, Wang S. A Second Wave? What Do People Mean by COVID Waves? – A Working Definition of Epidemic Waves. Risk Management and Healthcare Policy. 2021;14:3775–3782. doi: 10.2147/RMHP.S326051 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Hale T, Angrist N, Hale AJ, Kira B, Majumdar S, Petherick A, et al. Government responses and COVID-19 deaths: Global evidence across multiple pandemic waves. PLOS ONE. 2021;16(7):e0253116. doi: 10.1371/journal.pone.0253116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Cortes C, Vapnik V. Support-vector networks. Machine Learning. 1995;20(3):273–297. doi: 10.1007/BF00994018 [DOI] [Google Scholar]
  • 22.Jović A, Brkić K, Bogunović N. A review of feature selection methods with applications. In: 2015 38th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO); 2015. p. 1200–1205.
  • 23. Li Y, Li T, Liu H. Recent advances in feature selection and its applications. Knowledge and Information Systems. 2017;53(3):551–577. doi: 10.1007/s10115-017-1059-8 [DOI] [Google Scholar]
  • 24. Abdulhussain SH, Mahmmod BM, Flusser J, AL-Utaibi KA, Sait SM. Fast Overlapping Block Processing Algorithm for Feature Extraction. Symmetry. 2022;14(4):715. doi: 10.3390/sym14040715 [DOI] [Google Scholar]
  • 25. Abdulhussain SH, Mahmmod BM, Baker T, Al-Jumeily D. Fast and accurate computation of high-order Tchebichef polynomials. Concurrency and Computation: Practice and Experience. 2022;34(27):e7311. doi: 10.1002/cpe.7311 [DOI] [Google Scholar]
  • 26. Mahmmod BM, Abdulhussain SH, Suk T, Hussain A. Fast Computation of Hahn Polynomials for High Order Moments. IEEE Access. 2022;10:48719–48732. doi: 10.1109/ACCESS.2022.3170893 [DOI] [Google Scholar]
  • 27.Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD’16. New York, NY, USA: Association for Computing Machinery; 2016. p. 785–794. Available from: 10.1145/2939672.2939785. [DOI]
  • 28.GitHub—fmfn/BayesianOptimization: A Python implementation of global optimization with Gaussian processes. Available from: https://github.com/fmfn/BayesianOptimization.

Decision Letter 0

Hilal Tayara

12 Dec 2022

PONE-D-22-27577COWAVE: A Labelled COVID-19 Wave Dataset for BuildingPredictive ModelsPLOS ONE

Dear Dr. Raman,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 26 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Hilal Tayara

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

3. Please ensure that you refer to Figures 8 and 9 in your text as, if accepted, production will need this reference to link the reader to the figure.

4. We note you have included a table to which you do not refer in the text of your manuscript. Please ensure that you refer to Tables 1, 6 and 8 in your text; if accepted, production will need this reference to link the reader to the Table.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript is interesting; however, the following comment should be addressed:

1: All the sections and subsections must be included in the text, Such as :( Data and Methods, Smoothing Algorithms, etc.).

2: The abstract needs to be improved to do justice to the main contributions of the paper, also it contains some abbreviations that need to be explained.

3: The contribution is not stated also add it at the end of the introduction section.

4: Please add future work to the conclusion section and discuss it briefly.

5: This study suffers from a fresh literature review. It is recommended to boost the literature review of this study.

6: Please add the system specifications used for the evaluation as well as the programming language.

7: There are some typos and grammatical errors that should be corrected.

Reviewer #2: Summary:

In this work, COVID-19 case data have been collected from around the world. The regions of waves are labelled. Also, XGBoost model is used to provide a minimum standard for future classifiers trained on this dataset. In addition, the utility of the dataset for the prediction of (future) waves.

The manuscript is interesting; however, the following comment should be addressed :

Abstract :

- - - - - - - - - - -

1 – Some facts about the collected dataset need to be included in the abstract.

Introduction Section :

- - - - - - - - - - - - - - - - - - - - - -

2 – The introduction need to be extended because it is very short.

3 – The contribution should be included as a list at the end of the introduction section .

Data and Methods Section :

- - - - - - - - - - - - - - - - - - - - - - - -

4 – This section is fine. No comments.

The Datasets Section :

- - - - - - - - - - - - - - - - - - - - - -

5 – More details about the dataset should be included .

6 – Visual analysis should be discussed thoroughly .

Results Section :

- - - - - - - - - - - - - - - - - - - - - -

7 – The author mentioned that supervised classifier can be used. Thus , the author need to refer to other types of feature extraction tools. For example: a) 10.1109/ACCESS.2022.3170893, b) 10.1002/cpe.7311, and c) 10.3390/sym14040715 .

Conclusion Section :

- - - - - - - - - - - - - - - - - - - - - -

8 – This section is fine. No comments .

General Comments:

- - - - - - - - - - - - - - - - -

9 - There are some grammatical errors that should be corrected . It is highly recommended to be proofed the manuscript carefully .

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Jul 25;18(7):e0284076. doi: 10.1371/journal.pone.0284076.r002

Author response to Decision Letter 0


10 Jan 2023

Response to Reviewer 1

1. All the sections and subsections must be included in the text, Such as :( Data and Methods, Smoothing Algorithms, etc.).

We have added all sections and subsections. Further, we have now (slightly) reorganized the sections and subsections for better readability.

2. The abstract needs to be improved to do justice to the main contributions of the paper, also it contains some abbreviations that need to be explained.

We have improved the abstract, with more details regarding the contributions made, the nature of the dataset created, and the baseline classifiers. All abbreviations used in the abstract, have also been explained.

3. The contribution is not stated also add it at the end of the introduction section.

We have added the contributions in the Introduction section (lines xx-yy).

4. This study suffers from a fresh literature review. It is recommended to boost the literature review of this study.

We have now included an additional paragraph in the “Introduction” section to boost the literature review, as per the reviewer’s suggestion (lines 8-13).

5. Please add the system specifications used for the evaluation as well as the programming language.

Since we are not reporting any runtimes, we have not reported system specifications. We have now clearly indicated the programming language, and that the codes were run on Google Colaboratory, at the beginning of the Data and Methods section (lines 28-32).

6. This study suffers from a fresh literature review. It is recommended to boost the literature review of this study.

We have now extended the Introduction section, with some methods of predicting COVID waves (lines 8-16).

7. There are some typos and grammatical errors that should be corrected.

We have carefully proofread the document again, and corrected all typos and grammatical errors that were found.

Response to Reviewer 2:

1. Some facts about the collected dataset need to be included in the abstract.

We have added the source of the initially collected dataset, in the abstract.

2. The introduction needs to be extended because it is very short.

We have now extended the Introduction section, with some methods of predicting COVID waves.

3. The contribution should be included as a list at the end of the introduction section.

We have included the contribution at the end of the introduction (lines 18-21).

4. This section is fine. No comments.

We thank the reviewer for their appreciation.

5. More details about the dataset should be included.

We have now included more details about the dataset used to build the COWAVE datasets.

6. Visual analysis should be discussed thoroughly.

We have improved the discussion of the visual analysis.

7. The author mentioned that a supervised classifier can be used. Thus, the author needs to refer to other types of feature extraction tools. For example: a) 10.1109/ACCESS.2022.3170893, b) 10.1002/cpe.7311, and c) 10.3390/sym14040715.

We have now cited other possible feature selection techniques (lines 223-224).

8. This section is fine. No comments.

We thank the reviewer for their appreciation.

9. There are some grammatical errors that should be corrected. It is highly recommended to be proofed the manuscript carefully.

We have carefully proofread the document again and corrected all typos and grammatical errors that were found.

Decision Letter 1

Hilal Tayara

24 Mar 2023

COWAVE: A Labelled COVID-19 Wave Dataset for Building

Predictive Models

PONE-D-22-27577R1

Dear Dr. Raman,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hilal Tayara

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Please address all comments raised by reviewr 4.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

Reviewer #4: (No Response)

Reviewer #5: All comments have been addressed

Reviewer #6: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: (No Response)

Reviewer #5: Yes

Reviewer #6: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: (No Response)

Reviewer #5: Yes

Reviewer #6: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: (No Response)

Reviewer #5: Yes

Reviewer #6: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: (No Response)

Reviewer #5: Yes

Reviewer #6: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: Summary:

In this work, COVID-19 case data have been collected from around the world. The regions of waves are labelled. Also, XGBoost model is used to provide a minimum standard for future classifiers trained on this dataset. In addition, the utility of the dataset for the prediction of (future) waves.

The authors have addressed the raised comments.

Comments :

Abstract :

- - - - - - - - - - -

1 – The abstract is fine. No further comments .

Introduction Section :

- - - - - - - - - - - - - - - - - - - - - -

2 – This section is fine. No further comments .

Data and Methods Section :

- - - - - - - - - - - - - - - - - - - - - - - -

3 – This section is fine. No further comments .

Results Section :

- - - - - - - - - - - - - - - - - - - - - -

4 – This section is fine. No further comments .

Conclusion Section :

- - - - - - - - - - - - - - - - - - - - - -

5 – This section is fine. No further comments .

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Reviewer #3: This version of the manuscript is well improved. The authors have addressed all reviewer comments. The manuscript can be accepted for publication.

Reviewer #4: - The abstract is long and NOT satisfactory. It should contain the following parts:

i. The importance of or motivation for the research.

ii. The issue/argument of the research.

iii. The methodology.

iv. The result/findings.

v. The implications of the result/findings.

-where is keyword list.Authors should add keyword list contain of 5 to 8 keywords.

-The motivation and contribution need to be improve in introduction.Author should also add seprate paragaraph of orgnization at the end of introducation.

-Was the dataset balanced? if the dataset is unbalanced and may affect the results significantly. The authors

should solve the problem of the unbalanced dataset.

-Clearly highlight the mathematical terms used in the paper and explain them in the text.

-Author should discuss more recent reference in introducation

*COVID-19 detection by dogs: From physiology to field application-a review article

*Combination of Angiotensin (1-7) Agonists and Convalescent Plasma as a New Strategy to Overcome Angiotensin Converting Enzyme 2 (ACE2) Inhibition for the Treatment of COVID-19

*Derivatization and combination therapy of current COVID-19 therapeutic agents: a review of mechanistic pathways, adverse effects, and binding sites

*The next frontier in vaccine safety and VAERS: Lessons from COVID-19 and ten recommendations for action

- Conclusion to be made more systematic and future scope to be elaborated more on technical features

that are planned to be added in the proposed system in the near future.

- The use of English language is fine, however, it is recommended to be checked once again.

Reviewer #5: The authors have addressed all the comments. The manuscript is well structured. Abstract is okay. data collection, Methodology, results and discussion are okay. Relevant articles were cited and properly referenced.

Reviewer #6: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: No

Reviewer #4: No

Reviewer #5: Yes: Boluwaji Ade Akinnuwesi

Reviewer #6: No

**********

Acceptance letter

Hilal Tayara

3 Apr 2023

PONE-D-22-27577R1

COWAVE: A Labelled COVID-19 Wave Dataset for Building Predictive Models

Dear Dr. Raman:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Hilal Tayara

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. LOWESS smoothed cases data for the Netherlands.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). Here, regions of different colours indicate different unique waves.

    (TIF)

    S2 Fig. LOWESS smoothed cases data for Egypt.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). Here, regions of different colours indicate different unique waves.

    (TIF)

    S3 Fig. LOWESS smoothed cases data for the Netherlands.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the green curve, smoothed data; the orange curve, the LOWESS smoothed doubling rate, and the red curve indicates whether the day is part of a wave.

    (TIF)

    S4 Fig. LOWESS smoothed cases data for Egypt.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the green curve, smoothed data; the orange curve, the LOWESS smoothed doubling rate, and the red curve indicates whether the day is part of a wave.

    (TIF)

    S5 Fig. LOWESS smoothed cases data for the Netherlands.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the orange curve, smoothed data, and the green curve indicates whether the day is part of a wave.

    (TIF)

    S6 Fig. LOWESS smoothed cases data for Egypt.

    The Y-axis shows the number of cases, for a given date. The X-axis ranges from 3-Jan-2020 (Day 0) to 20-Oct-2021 (Day 655). The blue curve represents the raw data; the orange curve, smoothed data, and the green curve indicates whether the day is part of a wave.

    (TIF)

    Data Availability Statement

    The data underlying the results presented in the study are available from the GitHub repository (https://github.com/RamanLab/COWAVE).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES