Abstract
Optimizing COVID-19 vaccine distribution can help plan around the limited production and distribution of vaccination, particularly in early stages. One of the main criteria for equitable vaccine distribution is predicting the geographic distribution of active virus at the time of vaccination. This research developed sequence-learning models to predict the behavior of the COVID-19 pandemic across the US, based on previously reported information. For this objective, we used two time-series datasets of confirmed COVID-19 cases and COVID-19 effective reproduction numbers from January 22, 2020 to November 26, 2020 for all states in the US. The datasets have 310 time-steps (days) and 50 features (US states). To avoid training the models for all states, we categorized US states on the basis of their similarity to previously reported COVID-19 behavior. For this purpose, we used an unsupervised self-organizing map to categorize all states of the US into four groups on the basis of the similarity of their effective reproduction numbers. After selecting a leading state (the state with earliest outbreaks) in each group, we developed deterministic and stochastic Long Short Term Memory (LSTM) and Mixture Density Network (MDN) models. We trained the models with data from each leading state to make predictions, then compared the models with a baseline linear regression model. We also remove seasonality and trends from a dataset of non-stationary COVID-19 cases to determine the effects on prediction. We showed that the deterministic LSTM model trained on the COVID-19 effective reproduction numbers outperforms other prediction methods.
Introduction
The supply of approved vaccines for the COVID-19 will be limited in early stages [1]. Therefore, the basic question of what the optimized vaccine distribution might be must be answered. Regarding this question, the Centers for Disease Control and Prevention (CDC) and the National Institutes of Health (NIH) asked the National Academies to perform a consensus study on the equitable allocation of the COVID-19 vaccines among potential recipients [2]. The study takes different factors into account, such as high-risk individuals, population health disparities, assuring communities of color about the vaccine, addressing vaccine hesitancy among individuals, and considering ethical values [3, 4]. However, one of the main criteria for optimizing vaccine distribution is the geographic spread of the active virus at the time of vaccination [2]. Therefore, predicting the future spreading patterns of virus across different regions is important.
In this article, we develop real-time approaches for predicting the behavior of COVID-19 in all US states. We use data from the Centers for Disease and Prevention website and create two time-series datasets of the number of confirmed cases, and the effective reproduction numbers for all US states. The effective reproduction number, Rt, is defined as “the average number of secondary cases of disease caused by a single infected individual over her or his infectious period” [5].
To avoid training the models for all states, we use a self-organizing map (SOM) [6] to categorize all states into four groups according to their similarity in the reported effective reproduction numbers. In each group, we select the leading state (the state with earliest outbreaks). A deterministic Long Short Term Memory (LSTM) model [7], recurrent neural network (RNN) model, and stochastic Mixture Density Network (MDN) model [8] are then trained on data from each of the leading states.
In the deterministic LSTM model, the network output is the number of confirmed cases and the value of effective reproduction number in the next time-step. We use an LSTM RNN because (1) more confirmed cases can lead to more potential infection among populations in the future, and therefore, retaining all relevant historical information is important, and (2) this intelligent sequence analysis model has been reported by several studies to have high efficiency in time series forecasting problems [9].
In the stochastic MDN model, the network output is parameters of mixture distributions rather than a direct prediction value. The proposed MDN model is a combination of LSTM layers and a mixture of distributions. In this model, LSTM layers supply parameters for one or several distributions, which are then combined with weighting [8]. Finally, a sample of data can be extracted from the developed mixture distributions as an actual prediction [10].
We then compare the performance of developed models with a baseline linear regression model [11]. We aim to study whether using deterministic and stochastic sequence-learning models might have better predictive performance than linear regression. We also use an Augmented Dickey Fuller test [12] to assess the stationary and non-stationary status of the input dataset. We then remove seasonality and trend from the non-stationary datasets to investigate their effects on predictive performance.
This article is structured as follows. Section two discusses a published article on using artificial intelligence and machine learning to predict the behavior of the COVID-19 pandemic. Section three presents a brief mathematical explanation of Rt, seasonal-trend decomposition, SOMs, RNNs, and mixture density networks (MDNs). Section four discusses the development of sequence learning predictive models. Finally, section five explains the experimental setup, performance metrics, and results.
Literature review
On December 8, 2019, the government of China reported treatment of several new virus cases of a disease later named coronavirus disease 2019 (COVID-19) [13]. Since then, COVID-19 has spread across many countries and become a pandemic. COVID-19 is a highly transmissible respiratory disease with symptoms such as cough, fever, and breathing problems; it spreads through contact with infected individuals [14]. In January 2020, the US reported its first confirmed case of COVID-19; in mid-February 2020, the COVID-19 pandemic began to cause unprecedented social and economic consequences [13]. On December 14, 2020, the CDC reported 16,113,148 confirmed COVID-19 cases and 298,266 deaths in the US [15]. In this dire situation, the successful prior application of artificial intelligence and machine learning in critical problems inspired researchers to use these techniques against the COVID-19 pandemic. Artificial intelligence and machine learning have been used in various areas of predicting, contact tracing, screening, forecasting, and drug development for the COVID-19 pandemic [16].
Ribeiro et al. [17] have used cumulative confirmed Brazilian COVID-19 cases to train a support vector regression algorithm to forecast case numbers 6 days in advance. Chakraborty and Ghosh [18] have developed a hybrid method based on a Wavelet-based forecasting model and autoregressive integrated moving average model to forecast case numbers 10 days in advance for France, India, Canada, South Korea, and the UK. Chakraborty and Ghosh [18] have indicated that these forecast numbers of COVID-19 cases can act as an early-warning for policymakers and can be useful for the efficient allocation of health care resources. Kapoor et al. [19] have used mobility data and Graph Neural Networks to predict COVID-19 cases and have reported a 6% lower root mean squared logarithmic error than the best-performing baseline models.
Hartono [20] has indicated that developing an efficient predictive model is difficult because of the unknown characteristics of the virus causing COVID-19, as well as the political and geographical influences. Hartono [20] has used a topological autoencoder (TA), a topological neural network, to map the transmission dynamics of COVID-19 spread in several countries. TA produces a two-dimensional map in which countries with similar transmission dynamics are located close to each other. After selection of a target location for forecasting, TA has been used to identify a reference location with similar transmission dynamics that experienced earlier spread of the virus causing COVID-19. Finally, LSTM has been trained on data from the reference location to forecast the COVID-19 distribution in the target location.
Tomar and Gupta [21] have used LSTM and curve fitting to predict the number of COVID-19 positive cases and the number of recovered cases in India 30 days in advance. In that study, the data were collected from January 30, 2020 to April 4, 2020; 80% of the data were used for training, and 20% were used for testing. Li et al. [22] have developed an integrated spatiotemporal model based on RNNs and epidemic differential equations to predict the number of COVID-19 cases in Italy 7 days in advance.
Arora et al. [9] have used RNN based LSTM variants including Deep LSTM, Bidirectional LSTM, and Convolutional LSTM to predict the number of COVID-19 cases in India 1 day and 1 week in advance. In that study, the states of India are categorized into different areas according to the daily growth rate and the number of confirmed COVID-19 cases. The dataset contains time-series data of confirmed COVID-19 cases from March 14, 2020 to May 14, 2020 for each state in India [9]. Arora et al. [9] have conducted an experiment on open source libraries and have used the Adam optimizer to optimize the mean squared error loss. The authors used the mean absolute percentage error (MAPE) to compare the performance of several predictive methods and found an average MAPE of 3.22% for bi-directional LSTM, 4.81% for Stacked LSTM, and 5.05% for conv-LSTM.
Shahid et al. [23] have used support vector regression, autoregressive integrated moving average, LSTM, and Bidirectional LSTM for predicting confirmed COVID-19 cases, deaths, and recoveries in Israel, Russia, Brazil, Spain, the UK, Germany, Italy, China, India, and the US. The study used the mean absolute error, root mean square error, and r2_score indices to measure the performance of the models. The methods were found to rank as follows from best performance to worst performance: Bidirectional LSTM, LSTM, support vector regression, and autoregressive integrated moving average.
Chimmula and Zhang [12] have collected data on the numbers of confirmed COVID-19 cases, of fatalities, and recovered patients in a time series format from the Canadian Health Authority and Johns Hopkins University. The Augmented Dickey Fuller test was used to identify the effects of trends on the dataset and to report the stationary and non-stationary nature of the data [12]. The study has also developed an LSTM model to forecast the pandemic outbreak in Canada.
Mathematical models
In this section, the mathematical formulae of effective reproduction numbers, SOMs, RNNs, and MDNs are explained.
Effective reproduction number
The effective reproduction number, Rt, is defined as “the expected number of new infections caused by an infectious individual in a population where some individuals may no longer be susceptible” [24]. One of the main reasons for calculating Rt is to determine how interventions and control efforts in population immunity, policy, and other elements affect transmission in specific time-steps [25]. Furthermore, Rt can be used to study real-time changes in COVID-19 transmission [24]. To bring the pandemic under control, Rt must be decreased to less than 1 and as close to 0 as possible [5]. Therefore, predicting Rt, which is situation- and time-specific, can aid in understanding the pathogen transmissibility during the COVID-19 pandemic in the future. Several methods have been developed to estimate Rt but we use the method of Cori et al. [5], in which the effective reproduction number is as follows:
(1) |
where It is the number of incidents of infections on day t, and ws is the generation interval, which is defined as “the time between the infection time of an infected person and the infection time of his or her infector” [26]. In this equation, the generation interval is the only parametric assumption adopted from Nishiura et al. [27]. That study obtained 28 infector-infectee pairs and used the log-normal distribution and the discretized gamma distributions to generate the results. Nishiura et al. [27] have reported the standard deviation and mean of the serial interval at 2.9 days (95% credible interval (CrI): 1.9, 4.9) and 4.7 days (95% CrI: 3.7, 6.0). For estimating Rt, the Excel file of EpiEstim package was borrowed from Cori et al. [5] (Please refer to https://github.com/RezaDavahli for input data; 10 February 2021) [28].
Seasonal-trend decomposition
Normally, time series data can be decomposed into the trend, seasonality, and residual, as represented in the following equation:
(2) |
where t = 1, 2, · · ·, N; xt is an original signal at time t; τt is the trend; st is the seasonality, which is the patterns that repeat with a period of time; and rt is the residual. Several decomposition algorithms have been proposed for periodic and non-periodic datasets [29]. In this article, we use Seasonal-Trend Decomposition in six steps, which have been fully discussed by Qin et al. [30].
Before removing the seasonality and trend, we apply the Dickey Fuller test to determine whether the datasets are stationary or non-stationary. For the stationary dataset, seasonality and trend are not removed.
Self-organizing map
Teuvo Kohonen developed the SOM as a new form of neural network architecture and learning algorithm in the 1980s [6]. SOM uses an unsupervised learning process to analyze and represent the basic structures of a dataset as a map [31]. Therefore, SOM is commonly used to convert high-dimensional datasets into one- or two-dimensional maps [32]. Suppose that the input variables are X = (x1,x2,⋯xp)′; the weight vector assigned to the node l is ul = (ul1,ul2,⋯ulp) ′; ulj is the weight associated with node l of input variable xj; and p is the number of input variables [33].
The learning concept of SOM involves detecting and moving the winning node closer to each training case. For this purpose, the Euclidean distance di between the weight vector and the input variables is calculated for each item i in the training case. Subsequently, the weights of the winning node with the smallest di are updated by a learning rule. In each step, the index q of the winning node is:
(3) |
where is the weight for the lth node on the sth step, αs is the learning rate for the sth step, and xi is the input variable for the ith training case. For the winner node, the update rule is:
(4) |
where uls+1 is set to uls for all non-winning nodes.
Recurrent neural networks
Deep learning methods are effective for prediction because they automatically extract appropriate features from datasets [34]. RNN, a deep learning method, can store extensive historical information and use it to accurately predict the next steps in time-series problems [35]. However, its main disadvantage is long training time, because of vanishing gradient problems [21]. To overcome this problem, the LSTM structure, comprising a cell, an input gate, an output gate, and a forget gate, was developed to consider a long-term dependency [7]. In this structure, the cell stores values over arbitrary time intervals, and the gates adjust the flow of information in the recurrent hidden layer, as represented in Fig 1 [21].
The states of an input gate, an output gate, and a forget gate can be demonstrated mathematically by five equations:
(5) |
(6) |
(7) |
(8) |
(9) |
(10) |
In these equations, σ is the logistic sigmoid activation function; Ct is the cell state; W indicates the weight matrices; and i, o, and f indicate the input gate, output gate, and forget gate, respectively [36]. In this structure, the input gate specifies the flow of information and protects the cell from irrelevant information, the forget gate deletes irrelevant information, and the output gate regulates the flow of information passing through the rest of the network [9].
Mixture density networks
MDNs are a combination of a neural network and a mixture of distributions, as represented in Fig 2. In MDNs, neural networks are used to model a mixture of components [37]. The main aspects of MDNs include the type of neural network, the number and size of the hidden layers, the dimension of the output, the number of input parameters, the type of distribution, and the number of distributions [37]. Unlike the LSTM deterministic model with fully determined outputs, MDNs estimate probability distributions of potential outcomes.
In the following equation, the mixture of the probability density function (PDF) p(x) is represented as a combination of the m PDFs with weights Ω = {ω0,…, ωm−1}, where the sum of weights is equal to 1:
(11) |
Each pj is a normal distribution defined by a variance σj and a mean μj, according to the following equation:
(12) |
The model can be fit to the following objective loss function:
(13) |
In this study, RNNs are used to output the parameters of a mixture model including the mixing coefficient of each Gaussian kernel (the probability of each kernel), and the mean and variance of each Gaussian kernel.
COVID-19 predictive models
In this section, the deterministic and stochastic sequence-learning models are explained. These models are used to predict the number of confirmed COVID-19 cases and the effective reproduction numbers in all states in the US. We use data from the Centers for Disease and Prevention website, and have developed a dataset of the number of confirmed COVID-19 cases in all states of the US from January 22, 2020, to November 26, 2020, as indicated in Table 1.
Table 1. The confirmed case dataset at one time-step.
Date | Alabama | Alaska | Arizona | Arkansas | California | Colorado | Connecticut | Delaware | Florida | … |
---|---|---|---|---|---|---|---|---|---|---|
3/29/2020 | 110 | 12 | 146 | 34 | 480 | 246 | 469 | 18 | 891 | … |
… |
Next, we use the EpiEstim package to compute effective reproduction numbers for all time-steps and all states, as represented in Table 2.
Table 2. The Rt dataset at one time-step.
Date | Alabama | Alaska | Arizona | Arkansas | California | Colorado | Connecticut | Delaware | Florida | … |
---|---|---|---|---|---|---|---|---|---|---|
3/29/2020 | 2.06 | 1.89 | 2.11 | 1.28 | 1.77 | 1.92 | 2.39 | 1.91 | 2.26 | … |
… |
Both datasets contain 310 rows (time-step-days) and 50 columns (US states). To decrease the dimensionality of datasets, we use SOM to categorize all states into four categories. We apply the Minisom package [39] to a dataset containing the effective reproduction numbers from August 26, 2020 to November 26, 2020 for all US states. In the dataset, time-steps are considered features, and states are nodes. We have categorized all states into four groups according to the behavior of the effective reproduction numbers over time, as represented in Fig 3.
As shown in Fig 3, most neighboring states are interestingly clustered into the same group, thus indicating that the COVID-19 behavior is similar in close states. This conclusion appears logical, because there is more commuting and traveling between neighboring states.
We also use the R package Chorddig [40] to visualize all relationships among states according to their similarities in effective reproduction number (Fig 4).
After categorizing the states into four groups, we select the state with the earliest outbreaks as the leading state in each group. These leading states are used for training the models. Two sequence-learning models are considered: a deterministic LSTM model and a stochastic LSTM/MDN model. Fig 5 represents the structure of the stochastic LSTM/MDN model.
In the stochastic LSTM/MDN model, the neurons corresponding to the means μk(x) are passed to the negative log likelihood cost, but neurons corresponding to the variances σk(x) are passed through an exponential function before moving to the negative log likelihood cost. To satisfy the constraint of a sum of weights equal to 1 (Ω = {ω0,…, ωm−1}), the neuron corresponding to weights passes through the softmax function. Softmax creates probabilities between 0 and 1 from real values that add up to 1:
(14) |
As described earlier, the probability density of yt can be calculated according to the following equation:
(15) |
where gk(yt|x) is represented in the following equation as the kth multivariate Gaussian kernel.
(16) |
where the vector μk(x) is the center of kth kernel. Finally, the error function is represented as follows:
(17) |
Both deterministic and stochastic models were trained to provide predictions for time-step t + 1 after input of values up to time-step t. However, the output of the LSTM model is a value, whereas the output of the LSTM/MDN model is a mixture density parameters of a Gaussian mixture distribution. Therefore, for the stochastic model, a sample selected from this Gaussian mixture distribution is considered a prediction of the next time-step.
Experimental study
In this section, the developed stochastic and deterministic models are evaluated on two datasets of confirmed COVID-19 cases and effective reproduction numbers (Please refer to https://github.com/RezaDavahli for models and input data; 10 February 2021). Then they are compared with a linear regression model to better understand their predictive ability. In the next experiment, after performing an Augmented Dickey Fuller test, we remove the seasonality and trend of the non-stationary dataset. We then investigate the performance of the developed models trained on the residuals dataset.
Experimental setup
The performance of the developed deterministic and stochastic models is evaluated with the datasets of confirmed COVID-19 cases and effective reproduction numbers. The datasets contain values from January 22, 2020 through November 26, 2020 (Please refer to https://github.com/RezaDavahli for models and input data; 10 February 2021). In each dataset, 95% of the data are used for training (including 76% for training and 19% for validation), and 5% are used for testing. The testing set is considered from November 11, 2020 to November 26, 2020. The number of days for the testing set was borrowed from Arora et al. [9] and Hartono [20] aiming to provide comparability of our results. For developing the training dataset, 14 previous days are used in one batch to train the model and predict the value for the next day (1 day in advance). The Tensorflow [41] and Keras [42] libraries are used for developing the networks. The list of parameters in the two models is shown in Table 3.
Table 3. List of parameters in the two models.
Elements | LSTM | LSTM/MDN |
---|---|---|
Time step length | Day | Day |
Normalization | Yes | Yes |
Number of sequences | 14 | 14 |
Number of hidden layers | 3 | 2 |
Number of nodes in each hidden layer | 50 | 10 |
Number of mixture Gaussian kernels | - | 1 |
Performance metrics
We use Mean Absolute Percentage Error (MAPE), which is the percentile error of the models, to test the performance of the developed predictive models [43]. As represented in the following equation, yi, t is the real value in state i at time-step t, whereas ŷi, t is the predicted value.
(18) |
We compare the developed stochastic and deterministic predictions with that of linear regression to better understand the performance of the models.
Performance results
To fully understand the efficient model, we report the average MAPE for all leading states and for different combinations of models and datasets, as shown in Figs 6 and 7.
Several specific patterns are seen among the data. First, the predictive models trained on effective reproduction numbers showed much better performance than models trained on confirmed cases. On average, there was a 16% difference between the predictions based on confirmed cases versus effective reproduction numbers. Second, unlike the confirmed cases dataset, the Rt dataset is stationary, and there is no need to remove the seasonality and trend. However, with the confirmed cases dataset, the greatest improvement in performance due to removal of seasonality and trend was seen in the stochastic LSTM/MDN model. Third, the deterministic LSTM model exhebited the best performance for the two datasets. The LSTM model trained on the effective reproduction number has the best performance, with 3.46% MAPE among all fusions.
We also represented the performance of models from November 11, 2020, to November 26, 2020 in the leading state of California in group one (see in Figs 8 and 9).
As shown in Fig 8, although deterministic LSTM had better performance, stochastic LSTM/MDN was more successful in following the trend of the actual data. However, stochastic LSTM/MDN was much more sensitive to large changes in the actual data. We also show the performance of models on COVID-19 datasets when seasonality and trend are removed in comparison to the original datasets in the leading state of California (Fig 9).
Limitations
In this study, we developed models to predict the behavior of COVID-19 within the leading US states. Therefore, the main limitation is that we did not consider the effect of states on one another. Many states issued a stay-at-home order, asking residents to stay at home, which reduced mobility between states.
In our subsequent study, we plan to investigate the impacts of mobility on the performance of the sequence learning models.
Although we indicated that the models trained on Rt have much better performance, there are some limitations associated with that. The main limitation is that Rt can be calculated using different methodologies, which do not give the same estimate. The final major limitation relates to using SOM for dividing US states into four groups. SOM uses an unsupervised learning process to analyze and represent the Rt dataset as a map. SOM decreased the dimensionality of the Rt dataset by clustering states based on similarities in their respective Rt numbers from August 26, 2020 to November 26, 2020. In the resulting map, most neighboring states were clustered together, but there were several exceptions. Because this is an unsupervised clustering technique, the reasoning behind the clusters and exceptions is not clear.
Conclusion
This study developed stochastic and deterministic sequence learning models based on RNNs and MDNs to predict the behavior of COVID-19 virus in different US states. We trained the models on historical confirmed cases and Rt patterns. The developed models can predict geographic spreading of the active virus. The primary dataset contains 310 time-steps and 50 features (US states). To avoid training the models for all states, we used the unsupervised learning methods of SOM to categorize all states into four groups according to their similarity in COVID-19 behavior. After selecting one state from each group as the leading state (the state with the earliest outbreak), we trained the developed models. We found that the predictive models trained on Rt have much better performance than those trained on confirmed cases. In addition, the deterministic LSTM model exhibited better performance than the stochastic LSTM/MDN and linear regression models. However, the stochastic model was more successful in predicting the trends in the actual dataset. Finally, LSTM trained on Rt showed the best performance, with a MAPE value of 3.46%.
Data Availability
All codes and datasets are held in a public repository: https://github.com/RezaDavahli.
Funding Statement
The author(s) received no specific funding for this work.
References
- 1.Persad G, Peek ME, Emanuel EJ. Fairly prioritizing groups for access to COVID-19 vaccines. Jama. 2020;324(16):1601–2. doi: 10.1001/jama.2020.18513 [DOI] [PubMed] [Google Scholar]
- 2.National Academy of Sciences. A Framework for Equitable Allocation of Vaccine for the Novel Coronavirus [Internet]. 2020 [cited 2020 Nov 30]. Available from: https://www.nationalacademies.org/our-work/a-framework-for-equitable-allocation-of-vaccine-for-the-novel-coronavirus
- 3.Abu-Raya B, Gantt S, Sadarangani M. Challenges in evaluating SARS-CoV-2 vaccines during the COVID-19 pandemic. CMAJ. 2020;192(34):E982–5. doi: 10.1503/cmaj.201237 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.DeRoo SS, Pudalov NJ, Fu LY. Planning for a COVID-19 Vaccination Program. JAMA. 2020;323(24):2458–9. doi: 10.1001/jama.2020.8711 [DOI] [PubMed] [Google Scholar]
- 5.Cori A, Ferguson NM, Fraser C, Cauchemez S. A New Framework and Software to Estimate Time-Varying Reproduction Numbers During Epidemics. Am J Epidemiol. 2013. Nov 1;178(9):1505–12. doi: 10.1093/aje/kwt133 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Kohonen T. The self-organizing map. Proc IEEE. 1990;78(9):1464–80. [Google Scholar]
- 7.Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. doi: 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
- 8.Bishop CM. Mixture density networks. Birmingham: Aston University;1994; [Google Scholar]
- 9.Arora P, Kumar H, Panigrahi BK. Prediction and analysis of COVID-19 positive cases using deep learning models: A descriptive case study of India. Chaos Solitons Fractals. 2020;139:110017. doi: 10.1016/j.chaos.2020.110017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Xu J, Rahmatizadeh R, Bölöni L, Turgut D. Real-time prediction of taxi demand using recurrent neural networks. IEEE Trans Intell Transp Syst. 2017;19(8):2572–81. [Google Scholar]
- 11.Ghosal S, Sengupta S, Majumder M, Sinha B. Linear Regression Analysis to predict the number of deaths in India due to SARS-CoV-2 at 6 weeks from day 0 (100 cases-March 14th 2020). Diabetes Metab Syndr Clin Res Rev. 2020;14(4):311–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Chimmula VKR, Zhang L. Time series forecasting of COVID-19 transmission in Canada using LSTM networks. Chaos Solitons Fractals. 2020;135:109864. doi: 10.1016/j.chaos.2020.109864 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Davahli MR, Karwowski W, Sonmez S, Apostolopoulos Y. The Hospitality Industry in the Face of the COVID-19 Pandemic: Current Topics and Research Methods. Int J Environ Res Public Health. 2020;17(20):7366. doi: 10.3390/ijerph17207366 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Yang Y, Zhang H, Chen X. Coronavirus pandemic and tourism: Dynamic stochastic general equilibrium modeling of infectious disease outbreak. Ann Tour Res. 2020;83:102913. doi: 10.1016/j.annals.2020.102913 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.CDC. COVID-19 Cases, Deaths, and Trends in the US | CDC COVID Data Tracker [Internet]. Centers for Disease Control and Prevention. 2020 [cited 2020 Dec 14]. Available from: https://covid.cdc.gov/covid-data-tracker
- 16.Lalmuanawma S, Hussain J, Chhakchhuak L. Applications of machine learning and artificial intelligence for Covid-19 (SARS-CoV-2) pandemic: A review. Chaos Solitons Fractals. 2020;139:110059. doi: 10.1016/j.chaos.2020.110059 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Ribeiro MHDM, da Silva RG, Mariani VC, dos Santos Coelho L. Short-term forecasting COVID-19 cumulative confirmed cases: Perspectives for Brazil. Chaos Solitons Fractals. 2020;135:109853. doi: 10.1016/j.chaos.2020.109853 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Chakraborty T, Ghosh I. Real-time forecasts and risk assessment of novel coronavirus (COVID-19) cases: A data-driven analysis. Chaos Solitons Fractals. 2020;135:109850. doi: 10.1016/j.chaos.2020.109850 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Kapoor A, Ben X, Liu L, Perozzi B, Barnes M, Blais M, et al. Examining covid-19 forecasting using spatio-temporal graph neural networks. ArXiv Prepr ArXiv200703113 [Internet]. 2020; Available from: https://arxiv.org/abs/2007.03113
- 20.Hartono P. Similarity maps and pairwise predictions for transmission dynamics of covid-19 with neural networks. Inform Med Unlocked. 2020;20:100386. doi: 10.1016/j.imu.2020.100386 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Tomar A, Gupta N. Prediction for the spread of COVID-19 in India and effectiveness of preventive measures. Sci Total Environ. 2020;728:138762. doi: 10.1016/j.scitotenv.2020.138762 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Li Z, Zheng Y, Xin J, Zhou G. A recurrent neural network and differential equation based spatiotemporal infectious disease model with application to covid-19. ArXiv Prepr ArXiv200710929 [Internet]. 2020; Available from: https://arxiv.org/abs/2007.10929
- 23.Shahid F, Zameer A, Muneeb M. Predictions for COVID-19 with deep learning models of LSTM, GRU and Bi-LSTM. Chaos Solitons Fractals. 2020;140:110212. doi: 10.1016/j.chaos.2020.110212 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Gostic KM, McGough L, Baskerville SA, Joshi K, Tedijanto C, Kahn R, et al. Practical considerations for measuring the effective reproductive number, Rt. medRxiv [Internet]. 2020; Available from: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7325187.1/ [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Sciré J, Nadeau SA, Vaughan TG, Gavin B, Fuchs S, Sommer J, et al. Reproductive number of the COVID-19 epidemic in Switzerland with a focus on the Cantons of Basel-Stadt and Basel-Landschaft. Swiss Med Wkly. 2020;150(19–20):w20271. doi: 10.4414/smw.2020.20271 [DOI] [PubMed] [Google Scholar]
- 26.Kenah E, Lipsitch M, Robins JM. Generation interval contraction and epidemic data analysis. Math Biosci. 2008. May;213(1):71–9. doi: 10.1016/j.mbs.2008.02.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Nishiura H, Linton NM, Akhmetzhanov AR. Serial interval of novel coronavirus (COVID-19) infections. Int J Infect Dis. 2020;93:284–6. doi: 10.1016/j.ijid.2020.02.060 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Davahli MR. Input Datasets, Developed Models, and Live Figures [Internet]. 2020. Available from: https://github.com/RezaDavahli [Google Scholar]
- 29.Gao J, Song X, Wen Q, Wang P, Sun L, Xu H. RobustTAD: Robust time series anomaly detection via decomposition and convolutional neural networks. ArXiv Prepr ArXiv200209545 [Internet]. 2020; Available from: https://arxiv.org/abs/2002.09545
- 30.Qin L, Li W, Li S. Effective passenger flow forecasting using STL and ESN based on two improvement strategies. Neurocomputing. 2019;356:244–56. [Google Scholar]
- 31.Kohonen T. Essentials of the self-organizing map. Neural Netw. 2013;37:52–65. doi: 10.1016/j.neunet.2012.09.018 [DOI] [PubMed] [Google Scholar]
- 32.Douzas G, Bacao F. Self-Organizing Map Oversampling (SOMO) for imbalanced data set learning. Expert Syst Appl. 2017;82:40–52. [Google Scholar]
- 33.Huang F, Yin K, Huang J, Gui L, Wang P. Landslide susceptibility mapping based on self-organizing-map network and extreme learning machine. Eng Geol. 2017;223:11–22. [Google Scholar]
- 34.Fiok K, Karwowski W, Gutierrez E, Reza-Davahli M. Comparing the quality and speed of sentence classification with modern language models. Appl Sci. 2020. Jan;10(10):3386. [Google Scholar]
- 35.Connor JT, Martin RD, Atlas LE. Recurrent neural networks and robust time series prediction. IEEE Trans Neural Netw. 1994;5(2):240–54. doi: 10.1109/72.279188 [DOI] [PubMed] [Google Scholar]
- 36.Colah’s blog. Understanding LSTM Networks [Internet]. 2015 [cited 2020 Dec 5]. Available from: http://colah.github.io/posts/2015-08-Understanding-LSTMs/
- 37.Davis CN, Hollingsworth TD, Caudron Q, Irvine MA. The use of mixture density networks in the emulation of complex epidemiological individual-based models. PLoS Comput Biol. 2020;16(3):e1006869. doi: 10.1371/journal.pcbi.1006869 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Borchers DO. A Hitchhiker’s Guide to Mixture Density Networks [Internet]. Medium. 2019 [cited 2020 Dec 5]. Available from: https://towardsdatascience.com/a-hitchhikers-guide-to-mixture-density-networks-76b435826cca
- 39.Vettigli G. MiniSom: minimalistic and NumPy-based implementation of the Self Organizing Map [Internet]. 2020 [cited 2020 Dec 12]. Available from: https://github.com/JustGlowing/minisom
- 40.Flor M. chorddiag: Interactive Chord Diagrams [Internet]. 2021 [cited 2020 Dec 12]. Available from: https://github.com/mattflor/chorddiag
- 41.Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, et al. TensorFlow: a system for large-scale machine learning. In: Proceedings of the 12th USENIX conference on Operating Systems Design and Implementation. USA: USENIX Association; 2016. p. 265–83.
- 42.Chollet F. Keras: The python deep learning library. ascl. 2018;ascl-1806. [Google Scholar]
- 43.De Myttenaere A, Golden B, Le Grand B, Rossi F. Mean absolute percentage error for regression models. Neurocomputing. 2016;192:38–48. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All codes and datasets are held in a public repository: https://github.com/RezaDavahli.