Skip to main content
PLOS One logoLink to PLOS One
. 2025 May 28;20(5):e0323787. doi: 10.1371/journal.pone.0323787

T-RippleGNN: Predicting traffic flow through ripple propagation with attentive graph neural networks

Anning Ji 1,#, Xintao Ma 2,*,#
Editor: Jinran Wu3
PMCID: PMC12118821  PMID: 40435165

Abstract

Recently, accurate traffic flow prediction has become a significant part of intelligent transportation systems, which can not only satisfy citizens’ travel need and life satisfaction, but also benefit urban traffic management and control. However, traffic forecasting remains highly challenging because of its complexity in both topology structure and time transformation. Inspired by the propagation idea of graph convolutional networks, we propose ripple-propagation-based attentive graph neural networks for traffic flow prediction (T-RippleGNN). Firstly, we adopt Ripple propagation to capture the topology structure of the traffic spatial model. Then, a GRU-based model is used to explore the traffic model through the timeline. Lastly, those two factors are combined and attention scores are assigned to differentiate their influences on the traffic flow prediction. Furthermore, we evaluate our approach with three real-world traffic datasets. The results show that our approach reduces the prediction errors by approximately 2.24%-62,93% compared with state-of-the-art baselines, and the effectiveness of T-RippleGNN in traffic forecasting is demonstrated.

1 Introduction

Traffic management and control have become more challenging with the rapid growth of the number of vehicles during the urbanization process. Effective analysis and prediction of dynamic traffic conditions play a crucial role in the planning and construction of new infrastructure and the management of intelligent transportation systems [1]. However, due to spatial dependence and temporal dependence, traffic flow prediction is becoming increasingly difficult, and traditional methods are not applicable.

Firstly, the change in traffic conditions is affected by the topology structure of the traffic. Roads are part of a complex web where congestion in one area can quickly impact surrounding regions owing to the interconnected nature of the roads [2]. When the traffic condition changes or congestion occurs at the upstream point of the road, other downstream points along the traffic flow direction would also be influenced by the transfer effect [3]. As shown in Fig 1, the volume of point A gradually affects the adjacent roads differently with the distance of neighbors. The impact moves further along the network from points B and C to D.

Fig 1. Spatial dependency is affected by the topology structure of traffic networks.

Fig 1

Due to the impact between adjacent roads, the short-term traffic flow is transferred from upstream A to downstream D from near to far.

Secondly, traffic flow prediction has the temporal characteristic, where the traffic condition changes dynamically over time, exhibiting long-term periodicity and trend. However, the uncertainties in the case of short periods, including the temporal evolution of the traffic flow caused by incidents, are difficult to forecast [4].

Traditional approaches generally deal with the temporal characteristic, and they are classified into two categories, namely statistical methods such as Auto-Regressive and Moving Average (ARMA)[5,6], Support Vector Regression (SVR)[7], and Kalman filtering method [8,9], and deep learning methods, like Deep Recurrent Neural Networks (RNN) [10] and its successors Long Short Term Memory (LSTM) networks [11]. However, these methods encounter problems when it comes to dynamic real traffic conditions because of their incapability of handling spatial information of roads. Therefore, graph neural networks (GNN) have lightened the researchers to mine the topology relation of the traffic roads [12], including T-GCN [2], GaAn [13], and traffic-GGNN [14]. The key idea of GNN is to explore the non-Euclidian correlations, especially the irregularity in the traffic topology.

These GNN-based traffic flow prediction (FTP) systems have proved to be effective and show superior performance because of the consideration of traffic topology. However, it is argued that two problems remain unsolved. On the one hand, GNN-based models capture the spatial relations by transforming and aggregating information through the edges of the road network, yet fail to model the dynamic characteristic of the traffic influence from one road to another. As a result, it is crucial to model the dynamic traffic propagation, especially the heavy traffic load impact, which transfers to neighbors from near to far. On the other hand, they normally apply GNN to explore the spatial information and then RNN for temporal information, aggregating those two characteristics with a multilayer perceptron (MLP) after two-aspect exploration is accomplished. We consider that this kind of fusion neglects the inner relation between spatial and temporal features, thus easily falling into partial optimum with respect to either spatial or temporal factors.

In order to overcome those limitations, we propose Ripplenet attentive graph neural networks to predict traffic flow (T-RippleGNN), which utilizes Ripple propagation to capture the traffic spatial information and the gated recurrent unit (GRU) to explore the traffic through the timeline. The key idea of traffic topology exploration of the GNN-based model is information propagation, with which the structure of the road network can be constructed. Inspired by Ripplenet[15], we believe that the traffic flow propagation is similar to actual ripples created by raindrops propagating on the water, specifically the heavy traffic volume could affect the neighbors gradually through the propagation. Secondly, a GRU-based module is used to extract the temporal correlations after each propagation step. Finally, those two factors are combined with attention mechanisms to assign different weights to different locations regarding spatial and temporal influence on traffic flow. Our codes are available at https://github.com/helenma27/TRippleGnn. In summary, the main contributions of this study are as follows:

  • The concept of Ripplenet is employed to mine the topology structure of traffic networks by propagation traffic flow conditions like ripples, which can effectively explore the high-order correlations among traffic roads.

  • The temporal and spatial features are integrated after each time step, which could explore the global context of traffic dynamic topology through the timeline. Then, we reweight the importance of the influences of road neighbors through the timeline.

  • Our framework is evaluated with extensive experiments on three real-world datasets, SZ-taxi, Los-loop, and PEMS-BAY. The results demonstrate that our framework achieves a better performance compared with other state-of-the-art algorithms.

The rest of this paper is organized as follows: firstly, some related works regarding traffic forecasting and GNN-based models are introduced in Section 2. Then, some notations are illustrated, and our framework is explained in Section 3. Following that, the experiments are demonstrated with the evaluation metrics, and the results are analyzed in Section 4. Finally, our work is concluded, and some future work is shown in Section 5.

2 Related work

Traffic forecasting has drawn increasing attention recently because of its fundamental applications in transportation aspects, including traffic planning and reducing traffic jams. However, efficient and accurate traffic forecasting remains challenging since the number of vehicles is continuously growing with the rapid urbanization process, and the dynamic pattern in traffic prediction from both spatial and temporal point of view.

Existing traffic forecasting can fall into two categories: knowledge-driven approaches and data-driven approaches. The former methods often need superior and professional knowledge, as they tend to precisely model the traffic system regarding traffic volume, speed, and density. The typical models include the queuing theory method and the “traffic velocity” method [16]. However, the traffic is not only impacted by the factors shown before, which is non-trivial to build a model that absorbs all factors.

The latter method uses historical traffic data to predict future traffic states, and they can also be classified into two groups: traditional methods and deep learning methods. The traditional methods involve statistical based (such as ARIMA [5] and vector autoregressive [17]) and machine learning based (such as K-nearest neighbor [18] and support vector machine [19]) ones. Yang et al. proposed a hybrid model that integrated the extreme learning machine and an improved whale optimizer to enhance the robustness of the prediction [20].However, the proliferation of data and the complexity of road networks relegate traditional methods to the background. Thus, the rise of deep learning methods has given traffic forecasting another possibility to improve the efficiency and accuracy of prediction. RNN [21,22] and its variants including LSTM[23] and GRU[24] can be directly used for traffic forecasting owing to their capability to efficiently deal with sequence information such as traffic data. Combined with attention mechanisms, RNN can adaptively select the most relevant hidden states in the history sequence to predict the current road states [2527]. For instance, Zheng et al. put forward Gman that integrates attention mechanisms with the encoder, capturing the non-linear correlations between different time slices [25]. TGN [28] provides a general framework that operates on dynamic graphs represented as event sequences and a novel training strategy that enables the model to learn from the sequential nature of the data while maintaining efficient parallel processing capabilities. Yang et al. pointed out that M-R-AR-BiGR [29] was combined with the traditional methods and bi-directional gate recurrent units to capture the temporal characteristics and improve the robustness.

Furthermore, the ability of graph neural networks (GNN) to capture the temporal-spatial dependency and represent the correlations for the non-Euclidean spaces has attracted the attention of researchers since 2019. It can be categorized into recurrent graph neural networks (RecGNN) [30], graph convolutional networks (GCN) [31], graph autoencoders (GAE)[32], and the Spatial-Temporal GNN [33]. Currently, the most widely used GNN for traffic forecasting is the GCN, which propagates the node states to the neighbors and then updates them, so that the whole graph topology is grasped [3436]. Additionally, recent research has introduced various GCN and RNN models that can effectively capture the complex interdependencies inherent in spatial and temporal data. T-GCN [3] integrates GNN with GRU to capture both the topology structure of the traffic network and traffic dynamic patterns. STFGNN [37] enhances traffic forecasting by fusing temporal and spatial graphs with GCN and gated convolutions, with the capability to effectively address long sequence data. DyHSL [38] improves the forecasting with hypergraphs using HGNN for dynamics and interactive convolutions for spatial-temporal relations, which can capture higher-order interactions but at the cost of increased computational complexity and scalability issues. STGNN [39] integrates a novel graph neural layer, a recurrent layer and a transformer layer to capture the temporal and spatial dependency, which is more feasible for long period traffic speed prediction.

However, our framework differs from the above literature since the propagation idea is followed to capture the traffic network topology gradually, especially in the heavy traffic states. Ripple propagation is combined with GRU in an iterative manner, where spatial propagation and temporal modeling are deeply intertwined. This enables T-RippleGNN to not only extract spatial features more effectively but also refine them iteratively based on temporal context, leading to a more comprehensive representation of traffic patterns. Both STFGNN and STGNN employ graph convolutional networks to model spatial dependencies, but their reliance on predefined static graphs limits their ability to adapt to dynamic traffic conditions and long-term prediction. In contrast, our model leverages Ripple to propagate traffic influences adaptively across multi-hop neighbors, capturing not only local spatial correlations but also global traffic patterns, especially under heavy traffic conditions.

3 Proposed work

Our framework T-RippleGNN is shown in Fig 2. The left part is the whole architecture of our framework, which contains spatial and temporal feature extraction, and a final attention mechanism to reweight the location influence. The traffic flow profile at each time step is taken as input. It is through each Ripple propagation and GRU module as shown at the right part of the figure. Then, the attention mechanism is displayed in the upper part of the figure, to differentiate the location weight according to both the spatial and the temporal influences on the final traffic flow prediction. One thing to mention is that the parameters of each GRU are shared through the whole framework.

Fig 2. The framework of T-RippleGNN: The right part represents the specific architecture of a RippleGNN cell, where Ripple means Ripple propagation for the input traffic conditions.

Fig 2

Then GRU is applied after Ripple propagation.

At first, some necessary notations and formulations are shown in subsection 3.1. Then, our framework module is explained by the modules, namely the Ripple module in subsection 3.2, the GRU in subsection 3.3, and prediction units in subsection 3.4. Finally, the learning algorithm will be discussed in subsection 3.5.

3.1 Formulation

First, some notations and formulations are introduced.

  • Traffic network: The traffic network can be regarded as graphs G=(V,E) with N traffic sensor nodes V={v1,v2,, vn} and edges eE between nodes (v1,v2)E. Although in reality, the roads are directed, taking into account that the traffic congestion is bi-directionally propagated to upstream and downstream roads, G is taken as an undirected graph, as the propagation is bi-directional [40]. The historical traffic conditions are denoted as X={xN1,xN2,, xNT}N×T of the traffic network. The traffic conditions can be traffic flow, speed, density, and so on.

  • Traffic flow prediction: Given the traffic network G=(V,E) and the historical traffic flow profiles X={xN1,xN2,, xNT}N×T, the traffic flow prediction problem is to learn a model f for calculation in the next time τ, Y^={Y^Nt+1,Y^Nt+2,, Y^Nt+τ}.

3.2 The Ripple module

In this module, the Ripple propagation is implemented to spread the traffic conditions over the connecting road nodes and to explore the spatial relevance of the traffic network. In this study, we argye that the traffic profiles can be propagated through the edges as the raindrops fall into the river and provoke ripples, in which multiple ripples superpose to form a resultant preference distribution of the traffic condition.

We first define the concept of the ripple set regarding to the traffic network as follows:

εnk={vt|(vh,e,vd)G,  vhεnk1} (1)

where k={1,2,,H}; H is the hop number, and εn0={n} refers to the initial point n. Therefore, the ripple set of a traffic node can be defined as triples with the start node to the connected node neighbors belonging to (k1)-hop relevant nodes.

Snk={(vh,e,vd)|(vh,e,vd)G,  vhεnk1} (2)

As shown in Fig 3, the traffic condition can be spread through the ripple set at each hop, and then propagates along the links in the traffic networks, from near to far. Considering the traffic congestion influence attenuated with the increase of the hop number k, as the ripple decays, Fig 3 also shows different grey-level with the decreasing relatedness between starting nodes with neighbors.

Fig 3. The water wave shape of Ripple propagation procedure: The start points εn0 are randomly selected.

Fig 3

The traffic conditions propagate to its immediate neighbors, forming the 1-hop ripple set (εn1). During this step, the feature embeddings of the seed nodes are transmitted to their neighbors. The strength of this influence is represented by the intensity of the circle’s color, with darker shades indicating stronger influence. This propagation continues hop-by-hop until the whole structure is propagated or the limit hop number is reached.

The propagation process can be illustrated as follows: Start points are randomly chosen as seeds εn0, and then the k-hop ripple set can be calculated by Eq. 2. Starting from the 1-hop ripple set εn1, we assign each neighbor a relevance probability that the traffic condition can be propagated from vh to vt through the edge eE as displayed in Eq.3.

pi=exp(vhTvdi)(vh, e,vd)Sn1exp(vhTvd) (3)

where s represents the embedding dimension, and vds. This probability pi is calculated based on the similarity between the embeddings of the seed node and its neighbors, revealing the impact strength between the connected traffic nodes.

Using the relevance probabilities, we aggregate the traffic conditions of the neighbors to compute the first hop response (on1) for the seed node, as displayed in Fig 3(b). This response represents the traffic state of the seed node after incorporating the influence of its immediate neighbors.

on1=(vhi, ei,vdi)Sn1pivdi (4)

Thus, the first response on1 of the traffic profile in node n can be obtained. The propagation continues iteratively for k-hops. At each step, the response from the previous hop (onk1) is used as the new seed to compute the next response (onk), by iteratively replacing vh by the previous response onk1 in Eq.3 with the node ripple sets Snh for k=1,2K. This allows the model to capture traffic influences from increasingly distant nodes, reflecting the dynamic and multi-scale nature of traffic networks. Then, the embedding of traffic node n can be considered as the combination of all responses:

on=on1+on2++onK (5)

To reduce the computational complexity, the hop number threshold is set to a small number, while the traffic condition can still be propagated to the whole network. The reason is that the relevant nodes or neighbours are still undercover as the ripples overlap during different hop propagations [15]. In other words, a traffic node can be reached from multiple paths in the traffic networks.

Therefore, by iteratively calculating, the traffic flow embedding vectors can be acquired and formalized as Ot={o1t,o2t,oNt}, where N denotes the number of traffic nodes.

3.3 The GRU module

To capture the temporal dependency, the GRU is adopted to deal with the time-sequence data. As shown in subsection 3.2, the traffic node embedding for the specific time step can be calculated by Ripple propagation. Ont is defined as the traffic condition of node n at time t. To corporate the spatial relations while processing the time sequence, we follow the work [14,41] that applies GRU to obtain the traffic flow embedding. Given the input obtained from the Ripple propagation Ot, and the hidden representations from the previous time step Ht, the operation of GRU can be expressed as follows:

zt=σz(WzOt+UzHt1+bz) (6)
rt=σr(WrOt+UrHt1+br) (7)
H^t=tanh(WhOt+Uh(rtH^t1)+bh) (8)
Ht=H^tzt+H^t1(1zt) (9)

where represents the element-wise multiplication and  σ is the sigmoid function; Wz,   Wr,  Wh, bz, br and bh are the trainable weights and bias of GRU. zt and rt are the update gate and the reset gate respectively. The reset gate determines how to combine new input information with previous memory, while the update gate defines the amount of previous memory saved to the current time step.

3.4 The prediction module

In this module, the importance of each node in the traffic condition propagation is reevaluated after the temporal characteristic extraction, which integrates the spatial and temporal dependency. We apply the self-attention mechanism which only takes the sequence Ht as input and calculates the attention score.

a=exp(WHT(HT)12)Nexp(WHT(HT)12) (10)
HoutT=aHT (11)

where W is the trainable parameter to balance the spatial influence inside Ht.

Then, HoutT can be fed into a multi-layer perceptron (MLP) to produce the final traffic prediction:

Y^=WpHoutT+bp (12)

where Wp and bp are the trainable weights and bias respectively.

3.5 Learning algorithm

The loss function of our framework is defined as follows:

L=LPRE+LREG (13)

RMSE is used as the prediction loss function.

LPRE=1si=1s(yiy^i)2 (14)

The last loss term is the regularization term that prevents overfitting, by taking all the parameters in the framework into consideration.

LREG=λ2W22 (15)

In conclusion, the whole learning algorithm is shown in Algorithm 1: The training process contains two procedures: from line 3–5, the spatial dependency is explored with Ripple propagation; and line 7–8 represents the training of GRU together with the attention unit:

Algorithm 1. The learning algorithm of T-RippleGNN.

Input: traffic condition  X={xN1,xN2,, xNT} , traffic graph G=(V,E) .

Output: Prediction Xpred={xNt+1,xNt+2,, xNt+τ}

1: Initialize the parameters, the initial node n

2: For i = 1,…, T do

3: For k hops do

4: Sample minibatch of positive and negative edges from G

5: Calculate the gradients on the minibatch by back-propagation according to Eq. 3–5

6: end for

7: Calculate the gradients on the minibatch by back-propagation according to Eq. 6–12

8: Update the parameters by gradient descent with learning rate η

9: end for

4 Experiments

In this section, the performance of our framework on three real-world datasets is shown, and the results with state-of-the-art baselines are compared, when parameter sensitivity experiments are conducted.

4.1 Dataset

Three common real-world datasets: SZ-taxi, Los-loop, and PEMS-BAY are implemented, and are widely used in the field of traffic prediction. The details are shown below:

  • SZ-taxi: This is the taxi trajectory of Shenzhen from Jan. 1 to Jan. 31, 2015. It contains 156 major roads of Luohu District as the study area. The speed of traffic on each road is calculated every 15 minutes.

  • Los-loop: This dataset is collected on the highway of Los Angeles County in real time by loop detectors. It includes 207 sensors, and its traffic speed is collected from 3/1/2012 to 3/7/2012. The traffic speed data is aggregated every 5 minutes.

  • PEMS-BAY: It contains 6 months of statistics on traffic speed, ranging from 1/1/2017 to 6/30/2017, including 325 sensors in the Bay area.

Totally 80% of the datasets are selected as the training set and the remaining as the test set. The detail statistics are shown in Table 1. The data of the previous 60 minutes is used to predict the speed for the next 15, 30, and 60 minutes.

Table 1. The details of the three real-world datasets.

Datasets SZ-taxi Los-loop PEMS-BAY
Nodes 156 207 325
Edges 532 2833 2369
Time Steps 2976 2016 52116
Mean 12.2 58.9 62.6

4.2 Baselines

Our results are compared with some state-of-the-art algorithms, which can be classified into traditional statistical methods, deep learning-based models, and GNN-based models. They are listed as follows:

ARIMA[5]: It is a widely-used in time series analysis, which is a classical model for predicting the future states in the timeline.

SVR[7]: It uses a support vector machine to process regression on the traffic data sequence, which is also a traditional statistical method.

FNN[42]: The Feed forward Neural network can be directly employed to deal with traffic sequences, but only model the temporal dependencies.

GRU[41]: It is RNN’s variant model, which can effectively deal with traffic sequences regarding temporal dependencies.

T-GCN[2]: Temporal GCN combines the graph convolutional network with the GRU for traffic forecasting.

STGNN[39]: It combines a recurrent layer and a Transformer layer for temporal dependency, and a Graph neural network layer to process the spatial dependency.

4.3 Experiment setup

Those models are evaluated with three evaluation metrics: RMSE, MAP, and MAPE, where the smaller values indicate better prediction performance.

  • (1)

    Root Mean Squared Error (RMSE):

RMSE=1Ni=1N(yiy^i)2

where N is the number of the samples; yi and y^i denote the actual traffic and the predicted values respectively.

  • (2)

    Mean Absolute Error (MAE):

MAE=1Ni=1N|yiy^i|

It is the average of the absolute errors.

  • (3)

    Mean absolute percentage error (MAPE):

MAPE=1Ni=1N|yiy^iyi|×100%

Besides, the ripple hop number of 3 is employed for SZ-taxi and Los-loop, and hop number of 2 is for PEMS-BAY. A larger number of hops hardly improves performance but results in heavier computational complexity. In addition, the embedding dimension and the learning rate are set as 32 and 0.001, respectively.

4.4 Results and discussions

In this subsection, the results of comparison with other baselines are presented, and then some parameter sensitivity experiments are discussed.

Overall comparison.

The comparison results with the baselines regarding the traffic prediction are illustrated in Table 2–Table 4. Then, the following observations can be obtained:

Table 2. Performance comparison of T-RippleGNN and other baseline methods on SZ-Taxi dataset.
Model 15min 30min 60min
RMSE MAE MAPE RMSE MAE MAPE RMSE MAE MAPE
ARIMA 6.804 4.680 29.73% 6.804 4.680 29.72% 6.796 4.676 29.71%
SVR 4.152 2.627 21.09% 4.168 2.689 21.83% 4.220 2.776 22.76%
FNN 6.534 3.347 29.29% 4.160 2.822 23.48% 4.201 2.887 22.89%
GRU 6.298 3.114 28.11% 4.098 2.898 22.58% 4.187 2.829 22.67%
T-GCN 4.137 2.825 23.00% 4.159 2.845 23.49% 4.200 2.802 23.25%
STGNN 4.072 2,734 22.16% 4.076 2.786 22.29% 4.087 2.770 22.83%
T-Ripple
GNN
4.041 2.658 21.79% 4.060 2.683 21.80% 4.069 2.740 22.12%
Table 3. Performance comparison of T-RippleGNN and other baseline methods on Los-loop dataset.
Model 15min 30min 60min
RMSE MAE MAPE RMSE MAE MAPE RMSE MAE MAPE
ARIMA 10.054 7.706 20.84% 10.059 7.711 20.86% 10.066 7.717 20.90%
SVR 5.651 3.864 9.33% 6.734 4.647 11.51% 8.027 4.320 14.65%
FNN 5.244 3.501 9.03% 6.148 3.780 9.88% 7.885 4.431 13.89%
GRU 5.217 3.344 8.41% 6.123 3.771 9.90% 7.466 4.415 13.06%
T-GCN 5.209 3.275 8.34% 6.117 3.734 9.76% 7.266 4.403 12.31%
STGNN 5.081 2.954 7.86% 6.034 3.489 9.26% 7.170 4.114 12.00%
T-Ripple
GNN
4.958 2.850 7.39% 6.001 3.382 9.14% 7.108 4.085 11.51%
Table 4. Performance comparison of T-RippleGNN and other baseline methods on PEMS-BAY dataset.
Model 15min 30min 60min
RMSE MAE MAPE RMSE MAE MAPE RMSE MAE MAPE
ARIMA 6.254 4.215 10.42% 6.592 3.899 8.16% 6.643 6.046 9.23%
SVR 4.905 4.530 7.57% 5.617 4.850 8.60% 6.610 5.520 10.10%
FNN 4.420 2.211 5.19% 4.632 2.391 5.43% 4.987 2.468 5.89%
GRU 4.192 2.054 4.85% 4.557 2.200 5.71% 4.962 2.374 5.70%
T-GCN 2.723 1.482 3.11% 3.314 1.777 3.70% 3.965 2.108 4.571%
STGNN 2.430 1.189 2.38% 3.271 1.466 3.09% 4.201 2.039 4.55%
T-Ripple
GNN
2.318 1.179 2.30% 3.017 1.424 3.05% 3.989 1.884 4.39%
  • Our framework outperforms other baseline models almost on all the long-term and short-term settings. For example, for the 15-minute traffic prediction on the PEMS-BAY dataset, compared with the best result of STGNN, the values of RMSE, MAE, and MAPE have been decreased by 2.24%, 3.52%, and 3.36% respectively. For the long-term prediction of 60 minutes, compared with the best result among baselines on the dataset PEMS-BAY, the performance is increased by 5.05%, 7.60%, and 3.52% for the three evaluation metrics respectively. The results demonstrate the effectiveness of our proposed model, especially on the dataset with complex topology. The reason is that Ripple propagation can further acquire inner correlation among roads, as the traffic information flows through the network. Moreover, our model uses a recurrent unit directly after each time propagation rather than a spatial-temporal separate extraction and a delayed fusion. Finally, an attention mechanism is applied to reweight the impact of different roads.

  • Traditional statistical methods including ARIMA and SVR perform well for a short-term prediction, especially for a 15-minute test on the SZ-Taxi. SVR outperforms the deep-learning methods and even our proposed model. However, their weakness lies in long-term prediction and complex road networks. For all the datasets, ARIMA and SVR have a higher RMSE and MAE compared with other RNN-based and GNN-based approaches, dealing with 45-minute and 60-minute predictions.

  • Recently proposed GNN-based models perform better than traditional deep learning methods such as FNN and GRU. They achieve noticeable improvements on the datasets SZ-Taxi and PEMS-BAY. One reason is that they adopt graph neural networks to process the spatial dependency. Combined with temporal information, they can discover more correlation among data, which is consistent with the previous observations [12,42]. STGNN has a better performance compared with T-GCN because it explores both global and local temporal dependencies, preventing the local minimum problem.

Parameter variation.

Firstly, our framework is evaluated on the parameter sensitivity by varying the hop number from 1 to 5. The results are shown in Figs 46. Obviously, our framework achieves the best when for the SZ-Taxi and Los-loop datasets, and for the PEMS-BAY dataset. The reason is that a higher hop number of Ripple propagation module introduces noise during training, especially when low-order connectivity already includes those neighbors also covered by high-order propagation. It is of note that with, the performance is the worst as the propagation has not taken place, thus the model containing only the information of the time line, which is almost equivalent to GRU.

Fig 4. Prediction metrics for SZ-taxi.

Fig 4

Fig 6. Prediction metrics for PEMS-BAY.

Fig 6

Fig 5. Prediction metrics for Los-loop.

Fig 5

Moreover, the impact of the higher hop number on the dataset PEMS-BAY influences worse than on the dataset SZ-taxi, as the more complex the graph is, more noise is introduced by overlapped propagation traces. For instance, when h=5, RMSE = 4.960 for 60-minute prediction, with an increase of 19.58% compared with h=5 on the dataset PEMS-BAY. However, when h=5, RMSE = 4.42 for 60-minute prediction, increasing by 8.62% compared with h=5 on the dataset SZ-taxi. Therefore, the hop number should be carefully chosen: In simpler graphs (e.g., SZ-Taxi), higher hop numbers (h>3) introduce less noise, as the propagation traces are less likely to overlap. However, in more complex graphs (e.g., PEMS-BAY), higher hop numbers lead to significant performance degradation due to overlapping propagation paths and increased noise. Thus, in such cases, h3 is a better parameter option.

The size of the traffic condition is varied in each hop to further investigate the robustness of Ripple propagation. The result is shown in Fig 7, and it can be seen that the best embedding size occurs at 32. The RMSE first decreases as the size becomes larger, and then increases when reaching the best result, since a size of 32 is enough to carry the information for propagating to neighbor nodes. In more generalized cases, the embedding size is 32, and it achieves the best performance in the large-scale dataset PEMS-BAY, striking a balance between expressiveness and computational efficiency [15].

Fig 7. RMSE on the dataset PEMS-BAY with different ripple sizes.

Fig 7

In summary, while T-RippleGNN demonstrates effective performance in traffic flow prediction, it also has some limitations. Although the multi-hop propagation mechanism is effective for capturing spatial dependencies, it incurs increased computational overheads when applied to datasets with extremely large graph sizes. Specifically, the iterative aggregation of h-hop neighbors leads to a time complexity proportional to O(k×E), where E is the number of edges. Therefore, the choice of hyper parameters, particularly the maximum hop number h, should be adapted according to the graph complexity. While larger h values enable the capture of long-range spatial dependencies, they also introduce noise from irrelevant distant nodes and increase the risk of overfitting. Our ablation studies above reveal that performance plateaus when h=3 for most tested scenarios, reaching a balance between computational cost and efficiency.

5 Conclusion

To conclude, in this study, we propose T-RippleGNN for traffic forecasting. It explores the spatial correlation by Ripple propagation, and then investigates the temporal characteristic by GRU. Then, an attention mechanism is applied to reevaluate the importance of locations regarding spatial-temporal fusion. For experiment, our framework is tested compared with the state-of-the-art baselines. The results indicate that our framework improves the prediction quality and efficiency.

For future work, we plan to investigate other types of traffic conditions, such as traffic density or dynamic events. Instead of fixed hop numbers, we would also develop a dynamic mechanism to adjust the hyper parameters based on the real-time traffic congestion. Moreover, we would also further analyze the exploration of temporal dependency, especially the influence of global and local information of traffic flow sequences.

Data Availability

The underlying data from this study can be found at https://github.com/helenma27/TRippleGnn.

Funding Statement

X. Ma has received the funding Jilin Provincial Department of Education Science and Technology Research Project(JJKH20240199KJ). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Huang H. Dynamic modeling of urban transportation networks and analysis of its travel behaviors. Chin J Manage. 2005;2(1):18–22. [Google Scholar]
  • 2.Dong C, Shao C, Cheng-Xiang Z, Meng M. Spatial and temporal characteristics for congested traffic on urban expressway. J Beijing Univ Technol. 2012;38(8):128–32. [Google Scholar]
  • 3.Zhao L, Song Y, Zhang C, Liu Y, Wang P, Lin T, et al. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Trans Intell Transport Syst. 2020;21(9):3848–58. doi: 10.1109/tits.2019.2935152 [DOI] [Google Scholar]
  • 4.Ye J, Sun L, Du B, Fu Y, Tong X, Xiong H. Co-Prediction of Multiple Transportation Demands Based on Deep Spatio-Temporal Neural Network. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM. 2019. 305–13. doi: 10.1145/3292500.3330887 [DOI] [Google Scholar]
  • 5.Ahmed M, Cook A. Analysis of freeway traffic time-series data by using Box-Jenkins techniques. Transp Res Rec. 1979;(722):1–9. [Google Scholar]
  • 6.Hamed MM, Al-Masaeid HR, Said ZMB. Short-term prediction of traffic volume in urban arterials. J Transp Eng. 1995;121(3):249–54. [Google Scholar]
  • 7.Smola AJ, Schölkopf B. A tutorial on support vector regression. Statistics and Computing. 2004;14(3):199–222. doi: 10.1023/b:stco.0000035301.49549.88 [DOI] [Google Scholar]
  • 8.Kumar SV. Traffic Flow Prediction using Kalman Filtering Technique. Procedia Engineering. 2017;187:582–7. doi: 10.1016/j.proeng.2017.04.417 [DOI] [Google Scholar]
  • 9.Liu H, Tian H, Li Y. Comparison of two new ARIMA-ANN and ARIMA-Kalman hybrid methods for wind speed prediction. Applied Energy. 2012;98:415–24. doi: 10.1016/j.apenergy.2012.04.001 [DOI] [Google Scholar]
  • 10.Connor JT, Martin RD, Atlas LE. Recurrent neural networks and robust time series prediction. IEEE Trans Neural Netw. 1994;5(2):240–54. doi: 10.1109/72.279188 [DOI] [PubMed] [Google Scholar]
  • 11.Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. doi: 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
  • 12.Wu Z, Pan S, Chen F, Long G, Zhang C, Philip SY. A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst. 2020. [DOI] [PubMed] [Google Scholar]
  • 13.Zhang J, Shi X, Xie J, Ma H, King I, Yeung D. Gaan: gated attention networks for learning on large and spatiotemporal graphs. In: 2018. doi: 10.48550/arXiv.1803.07294 [DOI] [Google Scholar]
  • 14.Li S, Cheng Y, Jin P, Ding F, Li Q, Ran B. A Feature-Based Approach to Large-Scale Freeway Congestion Detection Using Full Cellular Activity Data. IEEE Trans Intell Transport Syst. 2022;23(2):1323–31. doi: 10.1109/tits.2020.3024670 [DOI] [Google Scholar]
  • 15.Wang H, Zhang F, Wang J, Zhao M, Li W, Xie X, et al. Ripplenet: propagating user preferences on the knowledge graph for recommender systems. In: Proceedings of the 27th ACM International Conference on Information and Knowledge Management. ACM. 2018. 417–26. [Google Scholar]
  • 16.Qi W, Li LI, Jianming HU, Zou B. Traffic velocity distributions for different spacings. J Tsinghua Univ Sci Technol. 2011;51(3):309–12. [Google Scholar]
  • 17.Zivot E, Wang J. Vector autoregressive models for multivariate time series. Modeling financial time series with S-PLUS. Springer. 2006. p. 385–429. [Google Scholar]
  • 18.Zhang X, Guo-Guang H, Hua-Pu L. Short-term traffic flow forecasting based on K-nearest neighbors non-parametric regression. J Syst Eng. 2009;24(2):178–83. [Google Scholar]
  • 19.Jeong Y-S, Byon Y-J, Castro-Neto MM, Easa SM. Supervised Weighting-Online Learning Algorithm for Short-Term Traffic Flow Prediction. IEEE Trans Intell Transport Syst. 2013;14(4):1700–7. doi: 10.1109/tits.2013.2267735 [DOI] [Google Scholar]
  • 20.Yang Y, Tao Z, Qian C, Gao Y, Zhou H, Ding Z, et al. A hybrid robust system considering outliers for electric load series forecasting. Appl Intell. 2021;52(2):1630–52. doi: 10.1007/s10489-021-02473-5 [DOI] [Google Scholar]
  • 21.Li Y, Yu R, Shahabi C, Liu Y. Diffusion convolutional recurrent neural networks: data-driven traffic forecasting. In: Proceedings of the International Conference on Learning Representations. 2018. [Google Scholar]
  • 22.Anton-Păduraru D-T, Mindru DE, Stănescu RS, Trofin F, Cobuz C, Cobuz M, et al. Unraveling Metabolic Syndrome in Youth: The Obesity Epidemic’s Hidden Complication. Children (Basel). 2025;12(4):482. doi: 10.3390/children12040482 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Liu Y, Zheng H, Feng X, Chen Z. Short-term traffic flow prediction with Conv-LSTM. In: 2017 9th International Conference on Wireless Communications and Signal Processing (WCSP). IEEE. 2017. 1–6. doi: 10.1109/wcsp.2017.8171119 [DOI] [Google Scholar]
  • 24.Fu R, Zhang Z, Li L. Using LSTM and GRU neural network methods for traffic flow prediction. In: Proc 31st Youth Academic Annu Conf Chin Assoc Automat YAC. 2016. 324–8. [Google Scholar]
  • 25.Zheng C, Fan X, Wang C, Qi J. Gman: a graph multi-attention network for traffic prediction. In: Proceedings of the AAAI Conf Artif Intell. 2020. [Google Scholar]
  • 26.Guo S, Lin Y, Feng N, Song C, Wan H. Attention Based Spatial-Temporal Graph Convolutional Networks for Traffic Flow Forecasting. AAAI. 2019;33(01):922–9. doi: 10.1609/aaai.v33i01.3301922 [DOI] [Google Scholar]
  • 27.Yao H, Tang X, Wei H, Zheng G, Li Z. Revisiting Spatial-Temporal Similarity: A Deep Learning Framework for Traffic Prediction. AAAI. 2019;33(01):5668–75. doi: 10.1609/aaai.v33i01.33015668 [DOI] [Google Scholar]
  • 28.Rossi E, Chamberlain B, Frasca F. Temporal graph networks for deep learning on dynamic graphs. arXiv. 2020. doi: 10.48550/arXiv.2006.10637 [DOI] [Google Scholar]
  • 29.Yang Y, Gao Y, Wang Z, Li X, Zhou H, Wu J. Multiscale-integrated deep learning approaches for short-term load forecasting. Int J Mach Learn Cyber. 2024;15:6061–76. [Google Scholar]
  • 30.Jiang W, Luo J, He M, Gu W. Graph Neural Network for Traffic Forecasting: The Research Progress. IJGI. 2023;12(3):100. doi: 10.3390/ijgi12030100 [DOI] [Google Scholar]
  • 31.Henaff M, Bruna J, and LeCun Y. Deep convolutional networks on graph-structured data. arXiv preprint arXiv:1506.05163, 2015 [Google Scholar]
  • 32.Kipf TN, Welling M. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308, 2016. [Google Scholar]
  • 33.Li Y, Yu R, Shahabi C, Liu Y. Diffusion convolutional recurrent neural network: data driven traffic forecasting. In: ICLR. 2018. [Google Scholar]
  • 34.Hamilton WL, Ying Z, Leskovec J. Inductive Representation Learning on Large Graphs. In NeurIPS, 1025–35, 2017. [Google Scholar]
  • 35.Veličković P, Cucurull G, Casanova A, Romero A, Liò P, and Bengio Y. Graph attention networks. In ICLR,2018. [Google Scholar]
  • 36.Van den Berg R, Kipf T, and Welling M. Graph Convolutional Matrix Completion. In KDD, 2017. [Google Scholar]
  • 37.Li M, Zhu Z. Spatial-temporal fusion graph neural networks for traffic flow forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence. 2021. [Google Scholar]
  • 38.Zhao Y, Luo X, Ju W, Chen C, Hua X, Zhang M. Dynamic hypergraph structure learning for traffic flow forecasting. In: ICDE. 2023. [Google Scholar]
  • 39.Wang X, Ma Y, Wang Y, Jin W, Wang X, Tang C, et al. Traffic flow prediction via spatial temporal graph neural network. In: WWW ’20: Proceedings of the Web Conference 2020. 2020. [Google Scholar]
  • 40.Cui Z, Ke R, Wang Y. Deep stacked bidirectional and unidirectional LSTM recurrent neural network for network-wide traffic speed prediction. In: Proc 6th Int Workshop Urban Comput UrbComp. 2016. 1–9. [Google Scholar]
  • 41.Li Y, Zemel R, Brockschmidt M, Tarlow D. Gated graph sequence neural networks. In: Proc ICLR. 2016. 1–19. [Google Scholar]
  • 42.Sutskever I, Vinyals O, Le QV. Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems. 2014. 3104–12. [Google Scholar]

Decision Letter 0

Jinran Wu

14 Jan 2025

PONE-D-24-39892T-RippleGNN: Predicting traffic flow through Ripple propagation with attentive graph neural networksPLOS ONE

Dear Dr. Ma,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 28 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has spec6ific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Thank you for stating the following financial disclosure:  [Key Research Fund of Jilin Police College].

At this time, please address the following queries:

a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution.

b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

c) If any authors received a salary from any of your funders, please state which authors and which funders.

d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. In the online submission form, you indicated that [Our data are only available upon request.].

All PLOS journals now require all data underlying the findings described in their manuscript to be freely available to other researchers, either 1. In a public repository, 2. Within the manuscript itself, or 3. Uploaded as supplementary information.

This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If your data cannot be made publicly available for ethical or legal reasons (e.g., public availability would compromise patient privacy), please explain your reasons on resubmission and your exemption request will be escalated for approval.

5. Please ensure that you refer to Figures 5, 6 in your text as, if accepted, production will need this reference to link the reader to the figure.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript introduces a novel method for traffic flow prediction combining Ripple propagation with attentive graph neural networks. While the methodology is interesting and potentially impactful, some areas require clarification and tightening to enhance comprehensibility and academic rigor.

1. The abstract lacks precise quantification of the improvement over state-of-the-art methods. For instance, instead of ‘show the effectiveness of our approach’, provide concrete statistics.

2. The related work section briefly mentions GNNs but does not delve deeply into differences between T-RippleGNN and similar models like STFGNN and DyHSL. A more explicit comparison and explanation of advancements are necessary to contextualize the contribution.

3. The explanation of the Ripple module in Section 3.2 is overly dense and assumes a high level of familiarity with mathematical notations. Add an intuitive explanation or flowchart to complement the mathematical formulations, especially for concepts like Eq. (3) and Eq. (4).

4. The manuscript could benefit from a more transparent discussion of the limitations of T-RippleGNN. For example: What computational trade-offs exist for datasets with very large graph sizes? How does the choice of parameters (e.g., hop numbers) affect robustness?

5. Terms like "traffic profiles" and "traffic conditions" are used interchangeably. For clarity, standardize terms throughout the manuscript or explicitly define them early in the text.

6. While the analysis mentions hop numbers and embedding sizes, it does not explore how the choice of these parameters might generalize across different datasets. Consider including a discussion or additional experiments on parameter selection strategies for unseen scenarios.

7. The figures are informative but lack consistency in design (e.g., use of colors and legends). Ensure figures like Fig. 3 emphasize key trends or propagation mechanisms.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

PLoS One. 2025 May 28;20(5):e0323787. doi: 10.1371/journal.pone.0323787.r003

Author response to Decision Letter 1


9 Mar 2025

Response to Reviewer 1 Comments

Dear Reviewers,

We greatly appreciate the constructive comments from all the reviewers and the opportunity we were given to improve our paper.

Following your comments and advice, we carried out new experiments and analyses to clarify unclear parts and adding more details to make the paper more complete.

Below we provide our response to your questions and concerns. Our response follows the following format:

Your comments

Response: our response and revision summary

“… quoted text of major content changes …”

We would like to thank you again for giving us this opportunity to improve the manuscript.

Yours Sincerely,

Authors.

1. The abstract lacks precise quantification of the improvement over state-of-the-art methods. For instance, instead of ‘show the effectiveness of our approach’, provide concrete statistics.

Response:

Thank you for your advices, we have added concrete statistics in the abstract part to illustrate the improvement as follows:

We evaluate our approach with three real-world traffic datasets. The results show that our approach reduces the prediction errors by approximately 2.24%-62,93% compared with state-of-the-art baselines and demonstrates the effectiveness of T-RippleGNN in traffic forecasting.

2. The related work section briefly mentions GNNs but does not delve deeply into differences between T-RippleGNN and similar models like STFGNN and DyHSL. A more explicit comparison and explanation of advancements are necessary to contextualize the contribution.

Response:

Thank you for your suggestions. It would improves the contribution to briefly states the differences. Thus we add to parts to give more explicit comparisons.

First, we add an additional explanation for DyHSL, which may suffer the problem of overfit and higher computational cost, and then compared with a representative method called STGNN:

STFGNN [35] enhances traffic forecasting by fusing temporal and spatial graphs with GCN and gated convolutions, with the capability to effectively deal with long sequence data. DyHSL [36] improves the forecasting with hypergraphs using HGNN for dynamics and interactive convolutions for spatial-temporal relations, which can capture higher-order interactions but at the cost of increased computational complexity and scalability issues. STGNN [40] integrates a novel graph neural layer, a recurrent layer and a transformer layer to capture the temporal and spatial dependency, which is more feasible for long period traffic speed prediction.

Then we modify the last paragraph of the related work, and explains the difference of our models with others, emphasizing the importance of Ripple propagation:

However, our framework differs from the above literatures that we utilize the propagation idea to capture the traffic network topology gradually, especially the heavy traffic states. We combine Ripple propagation with GRU in an iterative manner, where spatial propagation and temporal modeling are deeply intertwined. This enables T-RippleGNN to not only extract spatial features more effectively but also refine them iteratively based on temporal context, leading to a more comprehensive representation of traffic patterns. To be specific, both STFGNN and STGNN employs graph convolutional networks to model spatial dependencies, but its reliance on predefined static graphs limits its ability to adapt to dynamic traffic conditions and long-term prediction. In contrast, our model leverages Ripple to propagate traffic influences adaptively across multi-hop neighbors, capturing not only local spatial correlations but also global traffic patterns, especially under heavy traffic conditions.

3. The explanation of the Ripple module in Section 3.2 is overly dense and assumes a high level of familiarity with mathematical notations. Add an intuitive explanation or flowchart to complement the mathematical formulations, especially for concepts like Eq. (3) and Eq. (4).

Response:

Thank you for your suggestions. First, we move Figure 3 to line 219 between the Eq.3 and Eq.4, which illustrate the procedure of ripple propagation. This placement ensures that readers can directly reference the graphical illustration while engaging with the textual description of the propagation process, thereby reinforcing their understanding of key concepts such as multi-hop ripple sets, relevance probability calculation (Eq. 3), and iterative traffic profile aggregation (Eq. 4).

Second, we rewrite the caption of Figure 3 to better explain the procedure as follows:

The water wave shape of Ripple propagation procedure. The start points are randomly selected. The traffic profiles propagate to its immediate neighbors, forming the 1-hop ripple set (). During this step, the feature embeddings of the seed nodes are transmitted to their neighbors. The strength of this influence is represented by the intensity of the circle's color, with darker shades indicating stronger influence. This propagation continues hop-by-hop until the whole structure is propagated or the limit hop number is reached.

Third, we add further explanation for Eq.3 and Eq.4, which explains the effect of each equation:

(3)

where is the embedding dimension, and . This probability is calculated based on the similarity between the embeddings of the seed node and its neighbors, which reveals the impact strength between the connected traffic nodes.

Using the relevance probabilities, we aggregate the traffic profiles of the neighbors to compute the first hop response () for the seed node, as shown in Figure 3(b). This response represents the traffic state of the seed node after incorporating the influence of its immediate neighbors.

(4)

Thus, we can obtain the first response of the traffic condition in node n. The propagation continues iteratively for k-hops. At each step, the response from the previous hop () is used as the new seed to compute the next response (), by iteratively replacing by the previous response in Eq.3 with the node ripple sets for This allows the model to capture traffic influences from increasingly distant nodes, reflecting the dynamic and multi-scale nature of traffic networks.

4. The manuscript could benefit from a more transparent discussion of the limitations of T-RippleGNN. For example: What computational trade-offs exist for datasets with very large graph sizes? How does the choice of parameters (e.g., hop numbers) affect robustness?

Response:

Your advices are valuable to improve the parameter variation test. We have addressed how the hop number affects the result from line 377 to line 389. Additionally, we add a discussion to explain the computational trade-offs in the large graph sizes from line 402 to 412 as follows:

In summary, while T-RippleGNN demonstrates effective performance in traffic flow prediction, also has some limitations. The multi-hop propagation mechanism although effective for capturing spatial dependencies, incurs increased computational overhead when applied to datasets with extremely large graph sizes. Specifically, the iterative aggregation of h-hop neighbors leads to a time complexity proportional to , where is the number of edges. Thus, the choice of hyper parameters, particularly the maximum hop number , should be adapted according to the graph complexity. While larger values enable the capture of long-range spatial dependencies, they also introduce noise from irrelevant distant nodes and increase the risk of overfitting. Our ablation studies above reveal that performance plateaus when for most tested scenarios, reaching a balance between computational cost and efficiency.

5. Terms like "traffic profiles" and "traffic conditions" are used interchangeably. For clarity, standardize terms throughout the manuscript or explicitly define them early in the text.

Response:

Sorry for the fuzziness of the terms. We have change all the terms to “traffic conditions” for clarity, because they mean the same thing. We have given the definition of “traffic conditions” in section 3.2.

6. While the analysis mentions hop numbers and embedding sizes, it does not explore how the choice of these parameters might generalize across different datasets. Consider including a discussion or additional experiments on parameter selection strategies for unseen scenarios.

Response:

We sincerely thank you for raising this important point regarding the generalization of parameter choices across different datasets. We agree that a deeper discussion on parameter selection strategies is crucial for ensuring the robustness and applicability of T-RippleGNN in unseen scenarios. Based on the reviewer’s suggestion, we have expanded our discussion in the revised manuscript to address this issue. Below, we summarize the key additions:

In Parameter Variation, we have added a detailed discussion on how the optimal hop number varies across datasets.

In simpler graphs (e.g., SZ-Taxi), higher hop numbers () introduce less noise, as the propagation traces are less likely to overlap. However, in more complex graphs (e.g., PEMS-BAY), higher hop numbers lead to significant performance degradation due to overlapping propagation paths and increased noise. Thus in such cases, is a better parameter option.

Our experiments demonstrate that an embedding size of 32 achieves the best performance including large-scale graphs like PEMS-BAY. This consistency suggests that the chosen embedding size strikes a balance between expressiveness and efficiency, making it suitable for a wide range of scenarios.

This is because a size of 32 is enough to carry the information for propagating to neighbor nodes. In more generalized cases, the embedding size of 32, which achieves the best performance in the large scale datasets PEMS-BAY, strikes a balance between expressiveness and computational efficiency[15].

In future work, we propose exploring adaptive hop selection mechanisms that dynamically adjusting hyper parameters based on graph properties such as node degree distribution or network diameter.

We believe these additions can strengthen the manuscript by providing a more comprehensive discussion of parameter selection and generalization. We thank you for their valuable feedback, which has helped us improve the clarity and depth of our work.

7. The figures are informative but lack consistency in design (e.g., use of colors and legends). Ensure figures like Fig. 3 emphasize key trends or propagation mechanisms.

Response:

Sorry for the inconvenience. We have chose to use orange, green and purple to be main color for data expression. Thus we change our figure 1 and 2.

We also change the legends of Figure 3 to illustrate the propagation mechanisms to be clearer.

Fig. 3. The water wave shape of Ripple propagation procedure. The start points are randomly selected. The traffic conditions propagate to its immediate neighbors, forming the 1-hop ripple set (). During this step, the feature embeddings of the seed nodes are transmitted to their neighbors. The strength of this influence is represented by the intensity of the circle's color, with darker shades indicating stronger influence. This propagation continues hop-by-hop until the whole structure is propagated or the limit hop number is reached.

Thank you again for helping us improving the clarity and depth of our work

Decision Letter 1

Jinran Wu

26 Mar 2025

PONE-D-24-39892R1T-RippleGNN: Predicting traffic flow through Ripple propagation with attentive graph neural networksPLOS ONE

Dear Dr. Ma,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by May 10 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: As noted in my previous review, the manuscript still contains issues with grammar and phrasing that affect clarity. While some improvements have been made, expressions like “traffic profiles can be floored through the edges” are unclear or incorrect. I recommend a thorough language revision to ensure technical accuracy and readability.

I suggest the authors consider citing the following recent works to enhance the related literature and better contextualize their contribution:

1. Multiscale-integrated deep learning approaches for short-term load forecasting

2. Temporal Graph Networks for Deep Learning on Dynamic Graphs

3. A hybrid robust system considering outliers for electric load series forecasting

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy .

Reviewer #1: No

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

PLoS One. 2025 May 28;20(5):e0323787. doi: 10.1371/journal.pone.0323787.r005

Author response to Decision Letter 2


11 Apr 2025

1. As noted in my previous review, the manuscript still contains issues with grammar and phrasing that affect clarity. While some improvements have been made, expressions like “traffic profiles can be floored through the edges” are unclear or incorrect. I recommend a thorough language revision to ensure technical accuracy and readability.

Response:

We apologize for the language problems in the manuscript. The language presentation was improved with assistance from a native English speaker with an appropriate research background. For example:

In this study, we argue that the traffic profiles can be propagated through the edges as the raindrops fall into the river and provoke ripples, in which multiple ripples superpose to form a resultant distribution of the traffic condition.

2. citing the following recent works to enhance the related literature and better contextualize their contribution.

Response:

Thank you for your suggestions. It would improves the related work for adding more literature.

First, we add the recommended literature in the reference part:

21. Y. Yang , Z Tao ,·Q Tao, Y Gao, J Wu. A hybrid robust system considering outliers for electric load series forecasting[J].Applied Intelligence, 2021.DOI:10.1007/s10489-021-02473-5.

29. E. Rossi , B Chamberlain, F Frasca. Temporal Graph Networks for Deep Learning on Dynamic Graphs[J]. 2020.DOI:10.48550/arXiv.2006.10637

30. Y. Yang, Y. Gao, Z. Wang, X. Li, H. Zhou, J. Wu. Multiscale-integrated deep learning approaches for short-term load forecasting. Int. J. Mach. Learn. & Cyber. 15, 6061–6076 (2024).

Then we contextualize the contribution in the relate work part:

Yang et al. proposed a hybrid model that integrated the extreme learning machine and an improved whale optimizer to enhance the robustness of the prediction [21].

TGN [29] provides a general framework that operates on dynamic graphs represented as event sequences and a novel training strategy that enables the model to learn from the sequential nature of the data while maintaining efficient parallel processing capabilities. Yang et al. pointed out that M-R-AR-BiGR [30] was combined with the traditional methods and bi-directional gate recurrent units to capture the temporal characteristics and improve the robustness.

Decision Letter 2

Jinran Wu

15 Apr 2025

T-RippleGNN: Predicting traffic flow through Ripple propagation with attentive graph neural networks

PONE-D-24-39892R2

Dear Dr. Ma,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jinran Wu, PhD

Academic Editor

PLOS ONE

Acceptance letter

Jinran Wu

PONE-D-24-39892R2

PLOS ONE

Dear Dr. Ma,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jinran Wu

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Availability Statement

    The underlying data from this study can be found at https://github.com/helenma27/TRippleGnn.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES