Abstract
COVID-19 has become a matter of serious concern over the last few years. It has adversely affected numerous people around the globe and has led to the loss of billions of dollars of business capital. In this paper, we propose a novel Spatial–Temporal Synchronous Graph Transformer network (STSGT) to capture the complex spatial and temporal dependency of the COVID-19 time series data and forecast the future status of an evolving pandemic. The layers of STSGT combine the graph convolution network (GCN) with the self-attention mechanism of transformers on a synchronous spatial–temporal graph to capture the dynamically changing pattern of the COVID time series. The spatial–temporal synchronous graph simultaneously captures the spatial and temporal dependencies between the vertices of the graph at a given and subsequent time-steps, which helps capture the heterogeneity in the time series and improve the forecasting accuracy. Our extensive experiments on two publicly available real-world COVID-19 time series datasets demonstrate that STSGT significantly outperforms state-of-the-art algorithms that were designed for spatial–temporal forecasting tasks. Specifically, on average over a 12-day horizon, we observe a potential improvement of 12.19% and 3.42% in Mean Absolute Error (MAE) over the next best algorithm while forecasting the daily infected and death cases respectively for the 50 states of US and Washington, D.C. Additionally, STSGT also outperformed others when forecasting the daily infected cases at the state level, e.g., for all the counties in the State of Michigan. The code and models are publicly available at https://github.com/soumbane/STSGT.
Keywords: Time-series forecasting, Spatial–temporal graphs, COVID-19 forecasting, Transformers
Graphical abstract
1. Introduction
The Coronavirus disease 2019 (COVID-19) is caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). The first known case was identified in Wuhan, China in December 2019, and it has since rapidly spread around the globe. The coronavirus causes infection to the upper respiratory tract which could prove fatal if the immune system does not respond (Harapan et al., 2020). The World Health Organization (WHO) declared COVID-19 as a global pandemic in March, 2020 (Cucinotta & Vanelli, 2020). Since then, COVID has disrupted many businesses and households worldwide. Millions of people around the globe got infected with the virus and numerous families lost their loved ones (Koh, Geller, & VanderWeele, 2021) due to complications that aroused from the virus. Small businesses have been the ones that were the most adversely affected due to COVID (Bartik et al., 2020). The global economy and financial markets were also severely affected due to COVID (Pak et al., 2020). This in turn affected a lot of people worldwide who lost their jobs due to the economic downturn.
Deep learning has revolutionized the areas of computer vision, natural language processing, robotics, medical image processing and many others. Deep learning removes the feature engineering process of traditional machine learning and learns in an end-to-end manner by automatically extracting meaningful and important patterns from the input data. For example, many researchers have used deep learning models for detecting COVID-19 from medical images such as X-rays (Awan et al., 2021, Minaee et al., 2020). However, such methods could only be applied for persons already affected by COVID and these are not ideal for prevention of COVID-19.
In such situations, the prevention of COVID-19 becomes a matter of serious concern as a lot of lives could be saved if we could prevent the disease. The most important question that arises regarding the prevention of COVID is whether we can forecast the future status of the disease. In other words, if we could forecast the total number of infected and death cases in advance, then a lot of precautions could be taken to stop the disease from spreading. This leads us to the problem of time series analysis of COVID cases, where we make use of the past historical COVID data to forecast the future COVID cases. It is also equally important to forecast the COVID cases over the long-term since this would provide some time for the health care workers to prepare for difficult situations and help them serve the infected patients in a more efficient way. Additionally, a long-term forecasting would enable the common people to take safety precautions and prevent the disease.
Graph Neural Networks (GNNs) are a special type of machine learning algorithms that considers a graph as input and learns to predict the attributes of the vertices and edges of the graph. Recently, GNNs have been used for numerous applications where the data could be modeled as a graph. GNNs could also be combined with time series analysis algorithms to construct the spatial–temporal graph neural network. One of the classic cases where a spatial–temporal graph neural network has been used is traffic forecasting. In traffic forecasting, many researchers have proposed complex spatial–temporal graph neural networks to accurately capture the complex pattern in the spatial–temporal time series data. However, there are very few research works that deal with time series forecasting with COVID data. Some researchers have used spatial–temporal graph neural networks to forecast the COVID cases (Kapoor et al., 2020) but such methods were applied during the start of the pandemic (summer 2020) and as such the size of the dataset was too small to draw an inference. Hence, it becomes critically important that we develop a deep learning model which uses the COVID data from the start of the pandemic to recent times, so that this model could be used for prevention of the disease more effectively.
In this paper, we develop a novel Spatial–Temporal Synchronous Graph Transformer network (STSGT) which uses the historical COVID data to forecast the future infected and death cases for all the US states. In STSGT, first we denote the US states as vertices of a graph and the daily infected or death cases as the vertex features for a given time-step. The physical distances between the states are used as an adjacency matrix. Then, a large spatial–temporal synchronous graph is generated from the historical time-steps, where each day denote a time-step. This novel process ensures that we can capture the complex non-linear spatial–temporal dependencies between adjacent time-steps in a synchronous fashion. We apply the multi-head self-attention mechanism of transformers (Vaswani et al., 2017) on the spatial–temporal synchronous graph to infer the dependency of a vertex of the graph with another vertex. After each vertex finds its importance to all other vertices based on attention scores, we use the graph convolution network (GCN) (Kipf & Welling, 2017) to aggregate the features (infected or death cases) from neighboring vertices and update the current vertex of the graph. Finally, we forecast the future COVID cases for both short-term, i.e., next day forecasting, and long-term, i.e., forecasting over the next 12 days.
The main contributions of our work are summarized as follows:
-
•
We propose a novel Spatial–Temporal Synchronous Graph Transformer network (STSGT) which can simultaneously capture the complex spatial and temporal dependencies of the COVID time series data and accurately forecast the future COVID cases by utilizing the multi-head self-attention mechanism of transformers combined with a graph convolution network (GCN).
-
•
To the best of our knowledge, this is the first work that performs such a thorough and detailed analysis with real world COVID-19 time series datasets over a long period of time from the start of the pandemic in March 2020 to Nov 2021 (right before the start of the Omicron wave).
-
•
Extensive experiments on two real-world public datasets demonstrate that our model STSGT could accurately forecast the daily infected and death cases at both the national (for all the US states) and state (e.g., for all the counties of the state of Michigan) levels. Our model significantly outperforms other state-of-the-art models designed for spatial–temporal forecasting tasks.
The rest of the paper is organized as follows: Section 2 reviews the relevant research works and their applications. Section 3 describes the details of our proposed model STSGT along with its components. Section 4 describes the experimental setup and the results of our experiments. Section 5 presents discussion and future work. Lastly, Section 6 presents our conclusion.
2. Related works
2.1. Graph Neural Networks
Graph Neural Networks (GNNs) have been significantly useful when the input data does not have a fixed structure as in an image or text and could be represented as a graph with a set of vertices, edges and an adjacency matrix. GNNs and its variants have been employed for numerous applications such as recommendation systems (Wu, Sun, Zhang, & Cui, 2020), protein interface prediction (Fout, Byrd, Shariat, & Ben-Hur, 2017) and semantic segmentation (Qi, Liao, Jia, Fidler, & Urtasun, 2017). For example, a combination of Convolution Neural Networks (CNNs) and graph relation networks have been used for brain connectivity network analysis (Banerjee et al., 2020).
Graph Convolution Network (GCN) (Kipf & Welling, 2017) was first developed in a transductive setting to apply the convolutional operation on graphs in the spectral domain. The attention mechanism was introduced in the spatial domain into the feature aggregation process of GCNs by Graph Attention Network (GAT) (Veličković et al., 2018), which learns the weights to aggregate features from the neighboring vertices. The first application of graph representation learning to large-scale graphs in an inductive setting was introduced by GraphSAGE (Hamilton, Ying, & Leskovec, 2017). Cluster-GCN (Chiang et al., 2019) removed the redundant computation in the neighborhood sampling of GraphSAGE by clustering the graph and then train a GCN. A major limitation of increasing the number of layers of GCN is that it leads to over-smoothing, where vertex representations become indistinguishable as the number of layers of the graph network increases (Li, Han, and Wu, 2018, Li et al., 2019, Rong et al., 2019, Xu et al., 2018, Zhao and Akoglu, 2019).
2.2. Spatial–Temporal forecasting
Spatial–Temporal forecasting is a fundamentally important problem and has lot of useful applications such as weather forecasting (Tekin, Karaahmetoglu, Ilhan, Balaban, & Kozat, 2021) and action recognition (Yan, Xiong, & Lin, 2018). In smart transportation, traffic forecasting is one of the most challenging problems where the goal is to forecast the future traffic status based on historical observations. Spatial–temporal graph neural networks have been extensively used for traffic forecasting. Spatio-temporal graph convolution networks (STGCN) (Yu, Yin, & Zhu, 2018) used a GCN and a gated 1-D CNN to capture spatial and temporal dependencies respectively. Diffusion convolutional recurrent neural networks (DCRNN) (Li, Yu, Shahabi, & Liu, 2018) used the diffusion convolution and the diffusion convolution gated recurrent unit to model the spatial and temporal dependencies respectively. An attention-based approach has been combined with GCN and used with the weekly, daily and hourly traffic data in ASTGCN (Guo, Lin, Feng, Song, & Wan, 2019). Graph WaveNet (Wu, Pan, Long, Jiang, & Zhang, 2019) uses a gated temporal convolutional network (TCN) (Lea, Vidal, Reiter, & Hager, 2016) with a dilation factor and a GCN for capturing the long-range temporal and spatial dependencies respectively for traffic forecasting. Spatial–temporal synchronous graph convolutional networks (STSGCN) (Song, Lin, Guo, & Wan, 2020) applied the GCN on a synchronous spatial–temporal graph to capture the complex spatial–temporal dependencies and heterogeneity in traffic data. Spatial–temporal fusion graph neural networks (STFGNN) (Li & Zhu, 2021) fused a temporal graph with the spatial graph to forecast the future traffic conditions. Spatial–temporal graph neural networks has also recently been used to forecast the COVID-19 cases (Kapoor et al., 2020).
2.3. Transformers
One of the most prominent architectures that have revolutionized the Natural Language Processing (NLP) field are transformers (Vaswani et al., 2017). Transformers introduced the multi-head self-attention mechanism which could capture the relationship between every pair of words in a sentence in a parallel manner, thereby improving over the performance of serial architectures such as long short-term memory (LSTM) (Hochreiter & Schmidhuber, 1997) and gated recurrent unit (GRU) (Chung, Gulcehre, Cho, & Bengio, 2014), which have the vanishing gradient problem (Hochreiter, 1998). Transformers and its recent variants such as BERT (Devlin, Chang, Lee, & Toutanova, 2018) and large language models such as GPT-3 (Brown et al., 2020) have been used for numerous NLP applications such as machine translation, sentiment classification, speech recognition and many others.
The self-attention mechanism of transformers have also been used in computer vision by vision transformers (Dosovitskiy et al., 2020) for image recognition at scale. Recently, the self-attention mechanism of transformers has been adopted to the spatial domain of GNNs with graphormer (Ying et al., 2021) and to the spatial–temporal forecasting domain with the spatial–temporal transformer network (STTN) (Xu et al., 2020).
3. Our approach
Our proposed Spatial–Temporal Synchronous Graph Transformer network (STSGT) could be summarized with three main points: (1) Generate the samples from the original time series data with a sliding window approach; (2) For each sample, connect each vertex with itself at the previous and the next time steps to construct a spatial–temporal synchronous graph; (3) Use STSGT layers to capture the complex spatial–temporal correlations and then forecast the future status of the spatial–temporal graph network.
In STSGT, we generate a large synchronous spatial–temporal graph with several days historical data and then calculate the multi-head self-attention between every pair of vertices of this large graph. This novel process not only captures the relationship between the COVID infected or death cases between neighboring states or counties in a given day but also simultaneously captures the relationship between those states or counties in the previous and following days. Note that STSGT is fundamentally different from existing approaches for spatial–temporal forecasting. First, in most other graph neural network-based spatial–temporal models, a GCN is used to capture the spatial dependency for a given time-step, and a separate 1D convolution unit or attention unit is used to capture the temporal dependency between the time-steps, while STSGT used the multi-head self-attentions to capture both spatial and temporal dependency synchronously. In the literature, the most similar model to our approach is STTN (Xu et al., 2020), which has separate spatial and temporal transformer units and hence cannot simultaneously capture the complex spatial–temporal dependencies between adjacent time steps. The synchronous mechanism in our model helps us to achieve better performance in forecasting.
In the following sections, we provide a detailed discussion on every component of STSGT.
3.1. Generate samples using a Sliding Window
We use a sliding window approach to generate the samples from the COVID-19 time series datasets as shown in Fig. 1.
Each of the time step comprises of a spatial graph , where and denote the set of vertices and edges of respectively. The number of vertices of the graph is denoted as , where . “” denotes the adjacency matrix of , where . For each vertex of the graph , there are certain features that are associated with it. Hence, we denote graph signal matrix for each time-step of the time series as , where .
Then, we can formulate the problem of COVID-19 time series forecasting as learning the function which uses the historical observations to forecast the future observations:
(1) |
where, and denote the forecasted and ground truth observations respectively.
3.2. Construction of the spatial–temporal synchronous graph
The goal behind the construction of the spatial–temporal synchronous graph is to capture the complex spatial–temporal dependencies that are present between vertices of previous and subsequent time-steps. To achieve this goal, we connected each vertex of the graph to itself in the previous and next time-step for the entire time period as illustrated in Fig. 2. This results in a larger spatial–temporal graph where a given vertex at a particular time-step can aggregate information from another neighborhood vertex of previous and subsequent time-steps. The larger spatial–temporal graph also helps in capturing the long-range dependencies between the vertices of different time-steps.
The adjacency matrix of the spatial graph with vertices at a given time-step is given as . Hence, the spatial–temporal synchronous graph that we construct will have its adjacency matrix , where . In this large graph , a given vertex’s index in the spatial graph will be mapped to a new index in the large spatial–temporal graph , and the new index will be: , where is the time-step number in the larger spatial–temporal synchronous graph .
In the large spatial–temporal synchronous graph , when a given vertex at time-step is connected to itself at subsequent time-step , the corresponding value in the large adjacency matrix with vertices is set to be 1. Hence, the adjacency matrix of is formulated as:
(2) |
where and denotes the same vertex and respectively in subsequent time-steps of the spatial–temporal synchronous graph .
We have shown the adjacency matrix of the spatial–temporal synchronous graph in Fig. 3. The diagonal of are the adjacency matrices of the corresponding spatial graphs for time-steps where . Hence, we can say that , where to for .
The off-diagonal matrices of indicate the connection between each vertex to itself in the previous or subsequent time-steps. Hence, we can say that the upper-triangular side of , i.e. for is formed as: , where , and , for and . In a similar manner, the lower triangular side of , i.e. for is symmetric to the upper triangular side for . All the other elements are zero as indicated in Fig. 3.
In our approach, the spatial adjacency matrix for graph at time-step is a weighted adjacency matrix formed by the physical distances between the vertices of which we denote as the states of US or the counties in a state. Hence, for this adjacency matrix , the feature aggregation process of GCN in our model’s STSGT layer will not be affected as each vertex of has a different influence on its neighboring vertices. But for the adjacency matrix of the large spatial–temporal synchronous graph , the off-diagonal matrices and () have entries that are . Our goal is to learn the temporal dependency between same vertices in adjacent time-steps ( and ) in an end-to-end fashion. This would make our feature aggregation process more general and flexible since the features between vertices of adjacent time-steps would be aggregated based on the learned temporal dependency. Hence, we define a mask matrix with a similar shape as , i.e. . We calculate the Hadamard-product between the mask matrix and the adjacency matrix to form an updated matrix as follows:
(3) |
where the mask matrix contains the learnable weights between the vertices of adjacent time-steps and during the training process, denotes the Hadamard-product and denotes the updated adjacency matrix. Subsequently, is used for all GCN operations on the spatial–temporal synchronous graph .
3.3. Spatial–temporal synchronous graph transformer network
The architecture of our proposed Spatial–Temporal Synchronous Transformer network (STSGT) is shown in Fig. 4.
The input of STSGT is generated from the graph signal matrix by considering continuous time-steps. Hence, we generate the spatial–temporal synchronous graph signal matrix which is basically the graph signal matrix of the large spatial–temporal synchronous graph .
The input graph signal matrix is passed through a single 1-D convolutional layer to map the input to a higher dimensional space. Hence, the output of the 1-D convolutional layer is the spatial–temporal synchronous graph signal matrix where , and then is passed on to the STSGT layers.
3.3.1. Spatial–Temporal Synchronous Graph Transformer layer
The architecture of the STSGT Layer is illustrated in Fig. 5.
Each layer comprises of the Spatial–Temporal Synchronous Transformer (STST) for calculating the self-attention between the vertices of the spatial–temporal synchronous graph . A conventional GCN follows the STST to aggregate the information obtained through self-attention mechanism, from neighboring vertices of all vertices of . A GCN layer on the output of STST is defined as:
(4) |
where is the output of STST, is the updated spatial–temporal synchronous adjacency matrix, is the learnable weight matrix, is the learnable bias, is the non-linearity such as ReLU applied after the graph convolution, and is the output of the GCN layer.
Additionally, residual connections are added as shown in Fig. 5 to allow the gradients to back-propagate to the initial layers of the model during the training process. Finally, the output of a STSGT layer is an updated spatial–temporal graph signal matrix which is passed on to the next STSGT layer.
3.3.2. Temporal and spatial encodings
Temporal and spatial encodings are added to the input spatial–temporal graph signal matrix , before passing it on to the first STSGT layer. This helps the model to distinguish between the different time-steps (temporal encodings) and the different vertices of a given time-step (spatial encodings). These temporal and spatial encodings are added as follows:
(5) |
where the graph signal matrix after temporal and spatial encoding is . The temporal encoding added is which basically assigns same values for all vertices of a given time-step and different values for vertices in different time-steps. The spatial encoding added is which basically assigns same values for a given vertex across all time-steps and different values for vertices in the same time-step .
3.3.3. Spatial–temporal synchronous self-attention
The architecture of the Spatial–Temporal Synchronous Transformer (STST) within each STSGT layer is illustrated in Fig. 6 (left). First, the layer normalization is applied to the input graph signal matrix . Subsequently, the output of the layer norm is passed to a multi-head spatial–temporal synchronous self-attention mechanism as illustrated in Fig. 6 (right). Layer normalization is applied again to the output of the multi-head self-attention. Finally, a three layer Multi-Layer Perceptron (MLP) is used to output the encoded graph signal matrix .
The architecture of each head of the spatial–temporal synchronous self-attention mechanism is illustrated in Fig. 6 (right). The input to each head is the spatial–temporal graph signal matrix . This matrix is reshaped to . Then is projected to three higher dimensional latent subspaces by generating the Query(Q), Key(K) and Value(V) matrices. The Q, K and V matrices are generated as follows:
(6) |
where, and , and , and . Here, and the projection matrices are feed forward neural networks.
In a similar manner as the encoder of the transformer (Vaswani et al., 2017), we calculate the similarity or attention score between the query and key matrices in the higher dimensional latent subspace. The attention score is calculated as:
(7) |
where, is the attention score obtained from self-attention between all the vertices of the spatial–temporal synchronous graph with graph signal matrix . and are the query and key matrices respectively, and is the dimension of the key matrix . The softmax function normalizes the scores and the scaling factor prevents the saturation due to the softmax function.
Finally, the self-attention scores are multiplied by the value matrix to update the features of each vertex of . The update is done as follows:
(8) |
where and are the attention score and value matrix respectively. The matrix obtained with Eq. (8) is passed through a fully connected layer to update the output spatial–temporal synchronous graph signal matrix .
3.3.4. Forecasting the future COVID status
We forecast the future values of the spatial–temporal COVID time series with an output layer (see Fig. 4) which consists of two 1D convolution layers to map the output of the last STSGT layer to the original dimension of the time series data. Let us denote the output of the last STSGT layer in Fig. 4 as , which is then transposed and reshaped to . We then use two 1D convolution layers on as follows:
(9) |
where, is output of our STSGT model, is the ReLU non-linearity, and are the learnable weights and bias of the first 1D convolution layer and and are the learnable weights and bias of the second 1D convolution layer respectively. Here, is the hidden layer dimension of the output layers and is the length of the forecasting horizon. Finally, the output is transposed to .
It should be noted that we forecast all the horizon time-steps at the same time instead of recursively forecasting one of the horizon time-step at a given time as done in previous works (Li, Yu, et al., 2018, Song et al., 2020, Yu et al., 2018).
4. Experiments
We evaluated our STSGT framework with three main tasks: (1) forecasting the COVID-19 daily infected cases for the 50 states of US and Washington, D.C.; (2) forecasting the COVID-19 daily infected cases at the state level with 83 counties for the state of Michigan; and (3) forecasting the COVID-19 daily death cases for the 50 states of US and Washington, D.C. We compared the performance of our proposed approach with baselines and state-of-the-art spatial–temporal forecasting algorithms. All our experiments were performed using the PyTorch 1.8.0 framework (Paszke et al., 2019).
4.1. Datasets
4.1.1. John Hopkins University (JHU) COVID-19 dataset
We used the publicly available COVID-19 Data Repository by the Center for Systems Science and Engineering (CSSE) at Johns Hopkins University (https://github.com/CSSEGISandData/COVID-19) (Dong, Du, & Gardner, 2020). Specifically, we used the time series dataset which is updated regularly with the number of infected and death cases across all states of the US. We used the time series data from March 15, 2020 to Nov 30, 2021 spanning 626 days.
4.1.2. New York Times (NYT) COVID-19 dataset
We used the publicly available data from The New York Times, based on reports from state and local health agencies (The New York Times (2021). Coronavirus (COVID-19) data in the united states, 2021). Similar to the JHU dataset, we used the time series data for the daily infected and death cases across all US states. We used the time series data from March 18, 2020 to Nov 30, 2021 spanning 623 days.
4.2. Comparison methods
4.2.1. Baseline method
We used the following baseline method for spatial–temporal forecasting, to compare with our proposed approach.
-
•
Auto Regressive Integrated Moving Average (ARIMA) (Li, Yu, et al., 2018) - ARIMA consists of an auto-regressive (AR) and a moving average (MA) component and is used to forecast future values of a stationary time series. For our experiments, we used an order of (5, 1, 0) for ARIMA. That is, we used 5 AR and 0 MA components with first order difference (d=1) on the input time series. We used ARIMA as a baseline comparison method since it is one of the most common algorithms for time-series analysis, and it has been extensively used as a comparison method by several state-of-the-art spatial–temporal forecasting algorithms.
4.2.2. State-of-the-art Spatial–temporal Forecasting Algorithms
We compared our proposed approach with the following state-of-the-art methods that were designed for spatial–temporal forecasting. For each of the models below, the number of features or channels for each vertex of the graph is 1, which is the daily infected or death cases. The number of vertices for the graph per time-step is 51 (50 US states and Washington, D.C.) and 83 (for Michigan). The history (M) and horizon (H) for the sliding window is selected as 12 for all the spatial–temporal forecasting algorithms below. The default values are used for the hyper-parameters in each model, which we detailed in their descriptions below.
-
•
Spatial–Temporal Graph Convolution Network (STGCN) (Yu et al., 2018) - STGCN uses the spatial–temporal convolution blocks to forecast the horizon. Each of its spatial–temporal blocks consists of spectral graph convolution to capture the spatial dependency between vertices of same time-step and an independent temporal gated convolution block to capture the temporal dependency between the time-steps. We used the STGCN implementation (https://github.com/FelixOpolka/STGCN-PyTorch) with 16 spatial and output channels for each STGCN block.
-
•
Attention-based Spatial–Temporal Graph Convolution Network (ASTGCN(r)) (Guo et al., 2019) - ASTGCN concatenates the output of several spatial–temporal blocks used independently on the hourly, daily and weekly spatial–temporal data to forecast the next hour status. Each of the spatial–temporal blocks consists of spectral graph convolution to capture the spatial dependency and a combination of spatial and temporal attention mechanism to capture the dynamic spatial and temporal correlations of the spatial–temporal data. We used only the recent component of ASTGCN (i.e., ASTGCN(r)), equivalent to only using the hourly data, to have a fair comparison with other methods. We used the ASTGCN implementation (https://github.com/guoshnBJTU/ASTGCN-r-pytorch) with 2 ASTGCN blocks, 16 time filters with stride 1 and 16 Chebyshev filters of order 2.
-
•
Graph WaveNet (Wu et al., 2019) - Graph WaveNet comprises of multiple layers with skip connections where each layer consists of two gated temporal convolutional networks (TCN) (Lea et al., 2016). A dilation factor is used for each TCN to consider the long-range temporal dependencies of the time series. The gated TCN is followed by an independent graph convolutional layer to capture the spatial dependencies. We used the Graph WaveNet implementation (https://github.com/nnzhan/Graph-WaveNet) with the “double transition” adjacency matrix. We used 2 blocks with each block having 2 layers. We used 16 residual and dilation channels along with 128 skip channels and 256 end channels. A kernel size was used for the filter and gate convolutions.
-
•
Spatial–Temporal Transformer Network (STTN) (Xu et al., 2020) - STTN has several spatial–temporal blocks, and each of these blocks consists of an independent spatial and temporal transformer. The spatial transformer combines self-attention with a GCN using a gated mechanism to capture the spatial dependency. The output of the spatial transformer is passed to the temporal transformer which uses the self-attention mechanism to capture the temporal dependencies. We used the STTN implementation (https://github.com/wubin5/STTN) with 2 layers of spatial and temporal transformer, embedding size of 16 for each vertex of the graph, a 400-dimensional temporal encoding for the temporal transformer, a Chebyshev polynomial of order 3 for the spatial GCN and 2 heads for the multi-head self-attention for both spatial and temporal transformer.
4.3. Evaluation metrics
In this paper, we use the following three metrics to evaluate the performance of all the models. The Mean Absolute Error (MAE), Root Mean Square Error (RMSE) and Root Mean Square Log Error (RMSLE) is defined as:
(10) |
(11) |
(12) |
where, denotes the natural logarithm.
In all of the above equations, is the total number of samples in the dataset, is the number of ground truth cases and is the forecasted number of cases. A lower MAE, RMSLE and RMSE indicate that the forecasted number of cases is closer to the ground truth number of cases.
4.4. Implementation details
For all our experiments, we used the past 12 days history (M=12) to forecast the next 12 days horizon (H=12) using the sliding window approach shown in Fig. 1. The rationale behind the choice of 12 days history and horizon is that most of the recent state-of-the-art spatial–temporal forecasting algorithms (Guo et al., 2019, Wu et al., 2019, Yu et al., 2018), use the previous 12 time-steps to forecast the next 3, 6, 9 or 12 time-steps. We generate the adjacency matrix for the spatial graph at a given time-step by the euclidean distances between the states of the US. This distance is calculated by considering the coordinates of each state which in turn is the average of the latitude and longitude of all the counties for a given state. After generating the adjacency matrix, we normalized it with the maximum value and used a maximum threshold value of 0.3 to generate the final adjacency matrix for all US states and the counties at the state level. The same adjacency matrix was used for both JHU and NYT datasets to forecast the daily infected and death cases for all 50 US states and Washington, D.C. A separate adjacency matrix was generated at the state level, e.g. the state of Michigan, to forecast the daily infected cases for all the counties.
We divided the entire dataset in chronological order with 80% training, 10% validation and 10% testing. In other words, we used the data from March 15, 2020 to July 28, 2021 as the training set, July 29 to Sept 28, 2021 as the validation set and Sept 29 to Nov 30, 2021 as the testing set. For the NYT dataset, March 18, 2020 was used as the start date as the cases for one of the states were not reported till this date. It should be noted that cross-validation cannot be performed since the validation dataset has to follow the chronological order. Z-score normalization was applied to the inputs, i.e., the train, validation and test datasets were separately normalized by subtracting the mean and dividing by the standard deviation. We used the Mean Absolute Error (MAE) as the loss function and trained all the models for epochs with early stopping by monitoring the MAE on the validation set. We clipped the gradients to and used a batch size of 16 for training all models with a learning rate of 0.001 without using any learning rate decay.
The performance of our proposed STSGT framework is dependent on several hyper-parameters. These hyper-parameters include the number of STSGT layers, the number of attention heads for multi-head attention in each STST layer, the dimension of the Query(Q), Key(K) and Value(V) matrices for calculating the self-attention within each attention head, and the length of the history time-steps (M) for 12 horizon time-steps (H=12) in the sliding window (see Fig. 1). We empirically decided these hyper-parameters such that they are consistent with other spatial–temporal forecasting algorithms and thus we can have a fair comparison. Specifically, we used 2 STSGT layers in our model. Within each STSGT layer, we used 2 STST+GCN layer combinations. For the self-attention, we selected the Query(Q), Key(K) and Value(V) dimensions to be 16 with 2 attention heads in each of the STST layers. Additionally, for the sliding window with 12 horizon time-steps (H=12), we choose 12 historical time-steps (M=12).
4.5. Experimental results
4.5.1. Forecasting JHU COVID-19 daily infected cases for US states
The results of using our proposed STSGT model and other models to forecast the daily infected cases for the 50 US states and Washington, D.C. using the JHU dataset are summarized in Table 1. We summed the daily infected cases for all counties of a given state to generate the ground truth number of infected cases for that state. Clearly, our proposed STSGT significantly outperformed all the other models by achieving the lowest MAE, RMSLE and RMSE for both short-term horizon (next day, ) and long-term horizon indicated by the mean over the next 12 days (). When comparing with the second best model ASTGCN(r), we observe that the mean MAE over the long-term, i.e. 12 days horizon, reduces by 131.22 while the RMSLE and RMSE reduces by 0.081 and 397.81 respectively. An interesting observation is that ASTGCN(r) and our proposed STSGT can better forecast the longer time horizons for all the states of US, compared to other models as seen by a lower 12 days mean MAE when compared to day 1 MAE.
Table 1.
Algorithm | Forecasting horizon |
MAE | RMSLE | RMSE |
---|---|---|---|---|
ARIMA (Li, Yu, et al., 2018) | Day 1 | 2204.42 | 5.996 | 3724.19 |
12 Days Mean | 2206.14 | 6.069 | 3895.20 | |
STGCN (Yu et al., 2018) | Day 1 | 1243.48 | 0.869 | 2518.78 |
12 Days Mean | 1296.86 | 0.905 | 2369.93 | |
ASTGCN(r) (Guo et al., 2019) | Day 1 | 1078.06 | 0.822 | 2369.56 |
12 Days Mean | 1076.29 | 0.903 | 2102.68 | |
Graph WaveNet (Wu et al., 2019) |
Day 1 | 1140.02 | 0.917 | 2203.16 |
12 Days Mean | 1315.56 | 0.969 | 2263.27 | |
STTN (Xu et al., 2020) | Day 1 | 1172.34 | 0.948 | 2475.33 |
12 Days Mean | 1198.16 | 0.984 | 2191.91 | |
STSGT (ours) |
Day 1 | 978.390 | 0.802 | 2129.74 |
12 Days Mean | 945.070 | 0.822 | 1704.87 |
A comparison of the average infected cases over the next 12 days horizon for four of the most affected US states due to COVID, is shown in Fig. 7. We compared the average ground truth and forecasted average infected cases for the states of New York, Florida, California and Texas for a period of 12 days ranging from Nov 6–17, 2021, which is a time frame from our testing set. We use the past 12 days historical data ranging from Oct 25–Nov 5, 2021, to forecast the next 12 days horizon infected cases (Nov 6–17, 2021). In Fig. 7, we observe that the average number of forecasted infected cases with our proposed STSGT is much closer to the ground truth number of average infected cases for most of the states, when compared to state-of-the-art spatial–temporal forecasting methods.
4.5.2. Forecasting JHU COVID-19 daily death cases for US states
The results of using our proposed STSGT model and other models to forecast the daily death cases for the 50 US states and Washington, D.C. using the JHU dataset are summarized in Table 2. Similar to Table 1, we summed the daily death cases for all counties of a given state to generate the ground truth number of death cases for that state. Our proposed STSGT outperformed all the other models by achieving the lowest MAE for the short-term horizon (H=1) and the lowest MAE, RMSLE and RMSE for the long-term horizon indicated by the mean over the next 12 days (H=12). When comparing with the second best model ASTGCN(r), we observe that the mean MAE over the long-term, i.e. 12 days horizon, reduces by 0.73 while the RMSLE and RMSE reduces by 0.005 and 0.11 respectively. We observe that even though ASTGCN(r) achieves lower RMSLE and RMSE for forecasting the next day (day 1) death cases, our proposed model STSGT could better forecast the death cases over the long-term (12 days) as observed by a lower MAE, RMSLE and RMSE.
Table 2.
Algorithm | Forecasting horizon |
MAE | RMSLE | RMSE |
---|---|---|---|---|
ARIMA (Li, Yu, et al., 2018) | Day 1 | 42.62 | 2.753 | 74.89 |
12 Days Mean | 43.66 | 2.698 | 110.60 | |
STGCN (Yu et al., 2018) | Day 1 | 26.25 | 0.966 | 55.58 |
12 Days Mean | 25.23 | 0.952 | 54.15 | |
ASTGCN(r) (Guo et al., 2019) | Day 1 | 21.20 | 0.849 | 47.73 |
12 Days Mean | 21.36 | 0.873 | 47.25 | |
Graph WaveNet (Wu et al., 2019) |
Day 1 | 23.02 | 0.889 | 48.99 |
12 Days Mean | 25.03 | 0.922 | 51.16 | |
STTN (Xu et al., 2020) | Day 1 | 21.31 | 0.863 | 48.52 |
12 Days Mean | 22.05 | 0.892 | 48.23 | |
STSGT (ours) |
Day 1 | 20.73 | 0.855 | 48.30 |
12 Days Mean | 20.63 | 0.868 | 47.14 |
4.5.3. Forecasting JHU COVID-19 daily infected cases for Michigan
The results of using our proposed STSGT model and other models to forecast the daily infected cases at the state level, e.g., the state of Michigan, using the JHU dataset, are summarized in Table 3. In this case, the daily infected cases of the individual counties of Michigan served as the ground truth. We excluded the correctional facilities, counties denoted as “Out of Michigan” and unassigned counties, thus resulting in 83 counties in total. Hence, we used 83 vertices for the graph per time-step (i.e. for each day). As shown in Table 3, our proposed STSGT significantly outperformed all the other models by achieving the lowest MAE, RMSLE and RMSE for both short-term horizon (H=1) and long-term horizon indicated by the mean over the next 12 days (H=12). When comparing with the second best model STTN, we observe that the mean MAE over the long-term horizon (12 days) reduces by 9.86 while the RMSLE and RMSE reduces by 0.267 and 17.49 respectively. This experiment clearly demonstrates that our model could perform very well at the state level and hence our model could be used for forecasting infected and death cases for any US state.
Table 3.
Algorithm | Forecasting horizon |
MAE | RMSLE | RMSE |
---|---|---|---|---|
ARIMA (Li, Yu, et al., 2018) | Day 1 | 124.52 | 3.851 | 270.78 |
12 Days Mean | 142.46 | 4.006 | 322.06 | |
STGCN (Yu et al., 2018) | Day 1 | 48.30 | 0.739 | 95.99 |
12 Days Mean | 57.31 | 0.753 | 126.73 | |
ASTGCN(r) (Guo et al., 2019) | Day 1 | 47.50 | 0.772 | 97.14 |
12 Days Mean | 55.20 | 0.785 | 122.10 | |
Graph WaveNet (Wu et al., 2019) |
Day 1 | 41.34 | 0.658 | 87.97 |
12 Days Mean | 64.05 | 0.879 | 143.61 | |
STTN (Xu et al., 2020) | Day 1 | 42.50 | 0.812 | 81.37 |
12 Days Mean | 49.33 | 0.810 | 108.31 | |
STSGT (ours) |
Day 1 | 22.04 | 0.430 | 40.78 |
12 Days Mean | 39.47 | 0.543 | 90.82 |
The two most populated counties in Michigan are Wayne county with a population over 1.7 million and Oakland county with a population over 1.2 million. We plotted the ground truth and forecasted number of infected cases for Wayne and Oakland counties for a time frame in the testing set ranging from Oct 4–Nov 5, 2021 as shown in Fig. 8. We use the 12 days historical data to forecast the next day infected cases in Fig. 8. Specifically, we use the data from Sept 22–Oct 3 to forecast the infected cases for Oct 4, then the data from Sept 24–Oct 5 to forecast the infected cases for Oct 6 and so on. We removed the days when the ground truth infected cases were zero as these cases were not reported in the JHU dataset for Michigan. In Fig. 8, we compare STSGT with other spatial–temporal forecasting algorithms and clearly observe that the forecasted number of infected cases with STSGT is much closer to the ground truth number of infected cases.
4.5.4. Forecasting NYT COVID-19 daily infected cases for US states
The results of using our proposed STSGT model and other models to forecast the daily infected cases for the 50 US states and Washington, D.C. using the NYT dataset are summarized in Table 4. For the NYT dataset, the daily infected cases were reported for each state thereby eliminating the need for the county-wise summation process as done for the JHU dataset. Clearly, our proposed STSGT significantly outperformed all the other models by achieving the lowest MAE and RMSE for the short-term horizon (H=1) and the lowest MAE, RMSLE and RMSE for the long-term horizon indicated by the mean over the next 12 days (H=12). When comparing with the second best model ASTGCN(r), we observe that the mean MAE over the long-term horizon (12 days) reduces by 52.77 while the RMSLE and RMSE reduces by 0.010 and 136.08 respectively. This clearly shows the advantage of our model for forecasting over the long-term horizon.
Table 4.
Algorithm | Forecasting horizon |
MAE | RMSLE | RMSE |
---|---|---|---|---|
ARIMA (Li, Yu, et al., 2018) | Day 1 | 2152.48 | 5.935 | 3311.26 |
12 Days Mean | 2185.04 | 5.968 | 3676.43 | |
STGCN (Yu et al., 2018) | Day 1 | 1340.13 | 0.975 | 2303.10 |
12 Days Mean | 1384.35 | 0.997 | 2357.91 | |
ASTGCN(r) (Guo et al., 2019) | Day 1 | 939.840 | 0.852 | 1582.46 |
12 Days Mean | 976.420 | 0.879 | 1626.42 | |
Graph WaveNet (Wu et al., 2019) |
Day 1 | 960.840 | 0.854 | 1506.59 |
12 Days Mean | 1173.00 | 0.919 | 1905.55 | |
STTN (Xu et al., 2020) | Day 1 | 1125.88 | 0.978 | 1794.09 |
12 Days Mean | 1177.85 | 1.029 | 1854.19 | |
STSGT (ours) |
Day 1 | 903.590 | 0.864 | 1465.55 |
12 Days Mean | 923.650 | 0.869 | 1490.34 |
4.5.5. Forecasting NYT COVID-19 daily death cases for US states
The results of using our proposed STSGT model and other models to forecast the daily death cases for the 50 US states and Washington, D.C. using the NYT dataset are summarized in Table 5. Similar to Table 4, the daily death cases were reported for each state thereby eliminating the need for the county-wise summation process as done for the JHU dataset. Clearly, our proposed STSGT significantly outperformed all the other models by achieving the lowest MAE and RMSLE for the short-term horizon () and the lowest MAE, RMSLE and RMSE for the long-term horizon indicated by the mean over the next 12 days (). When comparing with the second best model ASTGCN(r) for the long-term (12 days mean), we observe that the mean MAE over 12 days horizon reduces by 2.02 while the RMSLE and RMSE reduces by 0.025 and 0.02 respectively.
Table 5.
Algorithm | Forecasting horizon |
MAE | RMSLE | RMSE |
---|---|---|---|---|
ARIMA (Li, Yu, et al., 2018) | Day 1 | 49.25 | 2.693 | 107.41 |
12 Days Mean | 49.36 | 2.661 | 138.96 | |
STGCN (Yu et al., 2018) | Day 1 | 28.86 | 0.949 | 67.85 |
12 Days Mean | 28.43 | 0.959 | 63.87 | |
ASTGCN(r) (Guo et al., 2019) | Day 1 | 25.90 | 0.944 | 62.07 |
12 Days Mean | 24.40 | 0.932 | 57.47 | |
Graph WaveNet (Wu et al., 2019) |
Day 1 | 24.73 | 0.921 | 60.07 |
12 Days Mean | 26.18 | 0.965 | 60.57 | |
STTN (Xu et al., 2020) | Day 1 | 29.85 | 1.091 | 69.55 |
12 Days Mean | 28.09 | 1.079 | 62.97 | |
STSGT (ours) |
Day 1 | 22.88 | 0.912 | 61.39 |
12 Days Mean | 22.38 | 0.907 | 57.45 |
5. Discussion and future work
The present study investigated the impact and utility of spatial–temporal forecasting for COVID-19 time series analysis. Our proposed novel STSGT could accurately forecast the daily infected and death cases across the 50 US states and Washington, D.C. STSGT also performed extremely well at the state level by accurately forecasting the daily infected cases for all the counties of Michigan. Hence, our model could help the common people and the government by taking necessary precautions so that COVID could be prevented.
It should be noted that we have selected mid-March (March 15 or 18, 2020) as the start date for our analysis as the pandemic started in almost all of the US states around that date, and there were significant number of infected and death cases after mid-March. We chose Nov 30, 2021 as the end date of our analysis as the Omicron variant of COVID started dominating after this date, and the number of daily cases spiked to a new high. Hence, the daily cases from around December 1, 2021 until the end of Omicron wave (around Feb 15, 2022) formed a different pattern than what we observed from the start of the pandemic. We believe that our model needs to be re-trained again with the Omicron wave’s new data pattern to achieve accurate forecasting, and we leave this for future work. In future, we also plan to incorporate an efficient self-attention mechanism to further accelerate the training process of STSGT while maintaining similar performance as our current approach.
6. Conclusion
In this paper, we proposed a novel model which can synchronously capture the complex spatial–temporal dependencies in the COVID data using the multi-head self-attention. We show through extensive experiments on two real-world public datasets that our model can outperform the existing models specifically designed for spatial–temporal forecasting. Our proposed model could have a significant impact on society by making the common people more aware and help the government eradicate this pandemic in the near future.
Declaration of Competing Interest
The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: Ming Dong and Weisong Shi reports financial support was provided by National Science Foundation.
Acknowledgment
This work was partially supported by National Science Foundation (Grant No. 2027251).
Data availability
I have shared the link to my data/code at the Attach File step. The links to the datasets and code is included in the manuscript file.
References
- Awan M.J., Bilal M.H., Yasin A., Nobanee H., Khan N.S., Zain A.M. Detection of COVID-19 in chest X-ray images: A big data enabled deep learning approach. International Journal of Environmental Research and Public Health. 2021;18(19):10147. doi: 10.3390/ijerph181910147. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Banerjee S., Dong M., Lee M.-H., O’Hara N., Juhasz C., Asano E., et al. Deep relational reasoning for the prediction of language impairment and postoperative seizure outcome using preoperative DWI connectome data of children with focal epilepsy. IEEE Transactions on Medical Imaging. 2020;40(3):793–804. doi: 10.1109/TMI.2020.3036933. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bartik A.W., Bertrand M., Cullen Z., Glaeser E.L., Luca M., Stanton C. The impact of COVID-19 on small business outcomes and expectations. Proceedings of the National Academy of Sciences. 2020;117(30):17656–17666. doi: 10.1073/pnas.2006991117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brown T., Mann B., Ryder N., Subbiah M., Kaplan J.D., Dhariwal P., et al. Language models are few-shot learners. Advances in Neural Information Processing Systems. 2020;33:1877–1901. [Google Scholar]
- Chiang, W.-L., Liu, X., Si, S., Li, Y., Bengio, S., & Hsieh, C.-J. (2019). Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining (pp. 257–266).
- Chung J., Gulcehre C., Cho K., Bengio Y. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555. [Google Scholar]
- Cucinotta D., Vanelli M. WHO declares COVID-19 a pandemic. Acta Biomedicine. 2020;91(1):157–160. doi: 10.23750/abm.v91i1.9397. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Devlin J., Chang M.-W., Lee K., Toutanova K. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805. [Google Scholar]
- Dong E., Du H., Gardner L. An interactive web-based dashboard to track COVID-19 in real time. The Lancet Infectious Diseases. 2020;20(5):533–534. doi: 10.1016/S1473-3099(20)30120-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dosovitskiy A., Beyer L., Kolesnikov A., Weissenborn D., Zhai X., Unterthiner T., et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929. [Google Scholar]
- Fout A., Byrd J., Shariat B., Ben-Hur A. Protein interface prediction using graph convolutional networks. Advances in Neural Information Processing Systems. 2017;30 [Google Scholar]
- Guo S., Lin Y., Feng N., Song C., Wan H. vol. 33. 2019. Attention based spatial-temporal graph convolutional networks for traffic flow forecasting; pp. 922–929. (Proceedings of the AAAI conference on artificial intelligence). [Google Scholar]
- Hamilton, W. L., Ying, R., & Leskovec, J. (2017). Inductive Representation Learning on Large Graphs. In Proceedings of the 31st international conference on neural information processing systems (pp. 1025–1035).
- Harapan H., Itoh N., Yufika A., Winardi W., Keam S., Te H., et al. Coronavirus disease 2019 (COVID-19): A literature review. Journal of Infection and Public Health. 2020;13(5):667–673. doi: 10.1016/j.jiph.2020.03.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hochreiter S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems. 1998;6(02):107–116. [Google Scholar]
- Hochreiter S., Schmidhuber J. Long short-term memory. Neural Computation. 1997;9(8):1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
- Kapoor A., Ben X., Liu L., Perozzi B., Barnes M., Blais M., et al. 2020. Examining covid-19 forecasting using spatio-temporal graph neural networks. arXiv preprint arXiv:2007.03113. [Google Scholar]
- Kipf, T. N., & Welling, M. (2017). Semi-supervised classification with graph convolutional networks. In International conference on learning representations.
- Koh H.K., Geller A.C., VanderWeele T.J. Deaths from COVID-19. JAMA. 2021;325(2):133–134. doi: 10.1001/jama.2020.25381. [DOI] [PubMed] [Google Scholar]
- Lea C., Vidal R., Reiter A., Hager G.D. European conference on computer vision. Springer; 2016. Temporal convolutional networks: A unified approach to action segmentation; pp. 47–54. [Google Scholar]
- Li Q., Han Z., Wu X.-M. 2018. Deeper insights into graph convolutional networks for semi-supervised learning. arXiv preprint arXiv:1801.07606. [Google Scholar]
- Li, G., Müller, M., Thabet, A., & Ghanem, B. (2019). DeepGCNs: Can GCNs Go As Deep As CNNs?. In 2019 IEEE/CVF international conference on computer vision (pp. 9266–9275).
- Li, Y., Yu, R., Shahabi, C., & Liu, Y. (2018). Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. In International conference on learning representations.
- Li M., Zhu Z. vol. 35. 2021. Spatial-temporal fusion graph neural networks for traffic flow forecasting; pp. 4189–4196. (Proceedings of the AAAI conference on artificial intelligence). [Google Scholar]
- Minaee S., Kafieh R., Sonka M., Yazdani S., Jamalipour Soufi G. Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning. Medical Image Analysis. 2020;65 doi: 10.1016/j.media.2020.101794. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pak A., Adegboye O.A., Adekunle A.I., Rahman K.M., McBryde E.S., Eisen D.P. Economic consequences of the COVID-19 outbreak: The need for epidemic preparedness. Frontiers in Public Health. 2020;8 doi: 10.3389/fpubh.2020.00241. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paszke A., Gross S., Massa F., Lerer A., Bradbury J., Chanan G., et al. In: Advances in neural information processing systems 32. Wallach H., Larochelle H., Beygelzimer A., d’Alché Buc F., Fox E., Garnett R., editors. Curran Associates, Inc.; 2019. Pytorch: An imperative style, high-performance deep learning library; pp. 8024–8035. [Google Scholar]
- Qi, X., Liao, R., Jia, J., Fidler, S., & Urtasun, R. (2017). 3d graph neural networks for rgbd semantic segmentation. In Proceedings of the IEEE international conference on computer vision (pp. 5199–5208).
- Rong, Y., Huang, W., Xu, T., & Huang, J. (2019). Dropedge: Towards deep graph convolutional networks on node classification. In International conference on learning representations.
- Song C., Lin Y., Guo S., Wan H. vol. 34. 2020. Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting; pp. 914–921. (Proceedings of the AAAI conference on artificial intelligence). [Google Scholar]
- Tekin S.F., Karaahmetoglu O., Ilhan F., Balaban I., Kozat S.S. 2021. Spatio-temporal weather forecasting and attention mechanism on convolutional lstms. arXiv preprint arXiv:2102.00696. [Google Scholar]
- 2021. The New York Times (2021). Coronavirus (COVID-19) data in the United States. Retrieved https://github.com/nytimes/covid-19-data. [03 April 2022] [Google Scholar]
- Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., et al. Attention is all you need. Advances in Neural Information Processing Systems. 2017;30 [Google Scholar]
- Veličković, P., et al. (2018). Graph Attention Networks. In International conference on learning representations.
- Wu, Z., Pan, S., Long, G., Jiang, J., & Zhang, C. (2019). Graph WaveNet for Deep Spatial-Temporal Graph Modeling. In Proceedings of the twenty-eighth international joint conference on artificial intelligence (pp. 1907–1913).
- Wu S., Sun F., Zhang W., Cui B. 2020. Graph neural networks in recommender systems: a survey. arXiv preprint arXiv:2011.02260. [Google Scholar]
- Xu M., Dai W., Liu C., Gao X., Lin W., Qi G.-J., et al. 2020. Spatial-temporal transformer networks for traffic flow forecasting. arXiv preprint arXiv:2001.02908. [Google Scholar]
- Xu K., Li C., Tian Y., Sonobe T., Kawarabayashi K.-i., Jegelka S. 2018. Representation learning on graphs with jumping knowledge networks. arXiv preprint arXiv:1806.03536. [Google Scholar]
- Yan, S., Xiong, Y., & Lin, D. (2018). Spatial temporal graph convolutional networks for skeleton-based action recognition. In Thirty-second AAAI conference on artificial intelligence.
- Ying C., Cai T., Luo S., Zheng S., Ke G., He D., et al. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems. 2021;34 [Google Scholar]
- Yu, B., Yin, H., & Zhu, Z. (2018). Spatio-temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. In Proceedings of the 27th international joint conference on artificial intelligence.
- Zhao L., Akoglu L. 2019. Pairnorm: Tackling oversmoothing in gnns. arXiv preprint arXiv:1909.12223. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
I have shared the link to my data/code at the Attach File step. The links to the datasets and code is included in the manuscript file.