Shin y Yoon - 2023 - Performance Evaluation of Building Blocks of Spati
Shin y Yoon - 2023 - Performance Evaluation of Building Blocks of Spati
Shin y Yoon - 2023 - Performance Evaluation of Building Blocks of Spati
ABSTRACT The traffic forecasting problem is a challenging task that requires spatial-temporal modeling
and gathers research interests from various domains. In recent years, spatial-temporal deep learning models
have improved the accuracy and scale of traffic forecasting. While hundreds of models have been suggested,
they share similar modules, or building blocks, which can be categorized into three temporal feature
extraction methods of recurrent neural networks, convolution, and self-attention and two spatial feature
extraction methods of convolutional graph neural networks (GNN) and attentional GNN. More importantly,
the models have been mostly evaluated for their entire architectures with limited efforts to characterize and
understand the performance of each category of building blocks. In this study, we conduct an extensive, multi-
faceted experiment to understand the influence of building block selection on traffic forecasting accuracy,
considering environmental characteristics and dataset distributions. Specifically, we implement six traffic
forecasting models using three temporal and two spatial building blocks. When we evaluate the models on
four datasets with diverse characteristics, the results show each building block demonstrates distinguishable
characteristics depending on study sites, prediction horizons, and traffic categories. The convolution models
demonstrate higher overall forecasting performance than other models, whereas self-attention models show
competitiveness in less frequent traffic categories, transition states, and the presence of outliers. Based on
the results, we also suggest an adaptive model evaluation framework for category-wise predictions of test
sets based on the performance of the models on validation sets. The results of this evaluation framework
demonstrate improved forecasting accuracy at most by 3.7% without further sophistication in existing model
architectures. The results enhance the utility of existing models and suggest guidelines for researchers
building traffic forecasting model architectures and for practitioners implementing these state-of-the-art
techniques in real-world applications.
INDEX TERMS Comparative study, deep learning, graph neural networks, spatial-temporal representation,
time-series prediction, traffic forecasting.
2023 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
136478 For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/ VOLUME 11, 2023
Y. Shin, Y. Yoon: Performance Evaluation of Building Blocks of Spatial-Temporal Deep Learning Models
and loop detectors and deep learning techniques to learn layer that modeled traffic flow as a diffusion process on
from abundance of data, traffic forecasting problems garnered a graph and compared its performance with ChebNet [76]
much research interests in recent years. on a traffic flow dataset. Cui et al. [15] suggested traffic
In traditional approaches to the traffic forecasting prob- graph convolution (TGC) and compared it with spectral
lem, conventional time series models such as autoregres- CNN [77] and ChebNet [76] in terms of the number of
sive integrated moving average (ARIMA) [9] and vector parameters, computation time, and ability to extract localized
autoregressive (VAR) [10] have gained popularity. Other features. In addition, they showed that the TGC model
data-driven machine learning algorithms such as support outperformed the spectral GCN-based models in overall
vector regression (SVR) [11] and k-nearest neighbor (kNN) performance. Although these studies provide comparative
[12] have also been utilized. Some other studies imple- studies between new and existing GNN layers, they only
mented simulation [13], [14] and physical modeling [8]. discuss the performance in terms of overall accuracy and
Although these approaches all demonstrated promising efficiency. In the temporal dimension, Reza et al. [70] present
results, their applications have had limitations in accuracy, the overall performance comparison between SVR, LSTM,
spatial-temporal range, or computation time. GRU, and transformer without consideration of spatial fea-
The recent surge of deep learning algorithms offered tures. Therefore, an investigation beyond overall performance
methods to fit a wide variety of functions with a larger to characterize each building block is necessary to understand
number of parameters while avoiding overfitting problems, and justify traffic forecasting model architecture.
and researchers have been able to leverage these advanced This study addresses this gap by conducting an extensive
techniques to capture the complex spatial and temporal and multi-faceted experiment to characterize the building
features of transportation networks in traffic forecasting blocks of spatial-temporal deep learning models for traffic
problems [3]. Recurrent neural networks (RNN) have gained forecasting. First, we define the five categories of the building
popularity in capturing temporal features with their intrinsic blocks through an extensive literature review. They are
ability to handle sequential data [15], [16], [17], [18], [19], RNN, convolution, and self-attention for temporal feature
[20], [21], [22], [23], [24], [25], [26], [27], [28], [29], extractions, and convolutional GNN and attentional GNN
[30], [31], [32], [33], [34], [35], [36], [37], [38]. The for spatial feature extractions. Subsequently, we implement
RNN methods, however, have suffered from the vanishing six traffic forecasting models, each incorporating distinct
gradient problem, and the convolution-based temporal feature combination of three temporal and two spatial building
extraction has been suggested to overcome this intrinsic blocks. To construct the models, we draw three models from
problem of RNN [39], [40], [41], [42], [43], [44], [45], previous literature, each representing a temporal building
[46], [47], [48], [49], [50], [51], [52], [53], [54], [55], block. Through replacement of spatial building blocks in
[56], [57], [58], [59], [60]. More recently, the self-attention selected models with GCN [78] and GAT [79], we assemble
mechanism [61] has demonstrated meaningful advances in six traffic forecasting models for the experiment. Finally,
traffic forecasting [22], [40], [41], [43], [46], [51], [58], [60], we evaluate the performance of the models on four real-world
[62], [63], [64], [65], [66], [67], [68], [69], [70], [71], [72]. datasets with diverse characteristics. In the experiment,
Convolutional neural networks (CNN) and graph neural we assess the influence of building block selections, and
networks (GNN) provide key spatial feature extraction analyze the performance across different traffic categories
capabilities. Although they constitute the pioneering efforts and presence of outliers.
to adopt deep learning architectures to traffic forecasting As the results, we find that the convolution and self-
[20], [23], [25], [42], [48], [49], [50], [54], CNN models attention-based models demonstrate advantages over the
have limitations in modeling the complex topology of the RNN-based counterparts in extracting temporal features for
underlying transportation networks. In contrast, GNN takes traffic forecasting. In the overall performance evaluation,
advantage of the node-link structure to incorporate the the convolution models tend to outperform the self-attention
underlying transportation network topology. By modeling models in overall performance. However, the self-attention
traffic sensors and road segments as graph nodes, the hidden models show a smaller performance discrepancy in perfor-
representation of a target node is learned by aggregating mances between 15-min and 60-min predictions, indicating
information from the neighboring nodes connected by edges a potential advantage in long-term forecasting. In addi-
[15], [16], [17], [21], [22], [24], [26], [27], [28], [29], [31], tion, the self-attention provides more accurate results
[32], [34], [35], [36], [37], [38], [40], [41], [43], [44], [45], in low-frequency traffic categories, and shows higher
[46], [47], [51], [52], [53], [55], [57], [58], [59], [60], [63], robustness against outliers than the convolution models.
[64], [66], [67], [68], [71], [72], [73], [74], [75]. Furthermore, we suggest the adaptive model evaluation
Despite the success of deep learning models in processing framework that flexibly selects models to conduct prediction
large datasets with high accuracy, efforts to understand based on the category-wise performance evaluation. Using
each of theses components, or building blocks, of the this framework, traffic predictions with higher accuracy
models are limited. For spatial feature extraction, Li et al. can be achieved without further sophistication in model
[16] proposed diffusion convolution, a convolutional GNN architectures. In summary, our main contribution is fourfold:
• Categorize the building blocks for spatial-temporal support vector machine (SVM), and Kalman filtering [18],
deep neural networks for traffic forecasting through an [30]. When traffic data were categorized into congestion
extensive literature review. The categorization includes levels, the LSTM model combined with the restricted
three temporal feature extraction methods - RNN, Boltzmann machine (RBM) showed at most 93.8% accuracy
convolution, and self-attention - and two spatial feature for congestion prediction tasks [19]. The sequence-to-
extraction methods - convolutional GNN, and attentional sequence framework has been adopted in many models for
GNN. multiple prediction horizons [16], [17], [21], [27], [31],
• Conduct an extensive, multi-faceted experiment using [32], [36]. In Bai et al. [33], a linear transformation layer
six traffic forecasting models each representing a was implemented to conduct multi-step traffic prediction.
distinct combination of three temporal and two spatial Wang et al. [34] suggested a model that utilized GRU to
building blocks on four different datasets. The overall produce aggregated spatial-temporal representations. Several
performance evaluation discovers building block pairs models have employed multiple layers of RNN [16], [21],
that generally yield higher accuracy: convolutional GNN [33], [35], whereas others have used the attention mechanism
& convolution and attentional GNN & self-attention. [27], [28], [36], [37] to capture the long-term relationship in
• Discover the characteristics of each building block. traffic data.
The convolution-based temporal feature extraction max- Another building block to extract temporal features is
imizes the performance gain in frequent traffic cate- convolution. In the absence of sequential computation,
gories, whereas the self-attention and attentional GNN convolution have been able to efficiently train the models
have increased robustness in infrequent conditions, such and overcome the vanishing gradient problem of RNNs.
as low-frequency traffic categories, traffic transitions, Originally suggested to process image data, earlier CNN
and the presence of outliers. approaches have processed traffic data into an image
• Propose an adaptive evaluation framework for traffic with each row and column representing each node of the
forecasting, which makes predictions using multiple transportation network and time step, respectively [49], [54].
models based on the performance on distinct traffic Although these models have demonstrated higher forecasting
categories. The framework increases the previous state- power than traditional machine learning algorithms and
of-the-art performance by 3.7% in a highway traffic vanilla LSTM, the CNN structure is limited as it represents
speed prediction task, without further sophistication in only 1D spatial complexity. To model time series more
previous model architectures. appropriately, temporal convolutions such as the gated 1D
The remaining paper is organized as follows. Section II causal convolution [45], [47], [48], [56], [59], [60] applied
investigates the literature on deep learning models in traffic convolution operation only along the temporal dimension.
forecasting studies. The preliminaries for this study and By limiting the usage of future information during the
definitions are in Section III. The methods and data are temporal feature extraction stage, causal convolutions have
explained in Section IV, along with the experimental setting. become applicable to traffic time-series modelling problems.
In Section V, we present the results and discussion of the The dilated causal convolution [87] that applies dilation to 1D
experiment. Finally, Section VI provides the conclusion and causal convolution to increase the reception field size with a
future study. limited number of layers has shown improved performance
[40], [41], [43], [44], [46], [51], [53], [55], [57], [58], [73].
II. LITERATURE REVIEW Recently, self-attention has also been widely adopted in
Deep learning models have proven effective in various traffic forecasting studies. Reza et al. [70] demonstrated
research fields such as image classification [80], object recog- the advantage of the transformer architecture over RNN
nition [81], and machine translation [82]. With their ability to models. To impose sequential information of traffic data,
process huge data and model non-linear relationships, deep self-attention have been implemented with various positional
learning has also become cutting-edge in traffic forecasting encoding methods. While the original Transformer [61]
studies. Following earlier works on stacked autoencoders [83] implemented the sinusoid to encode the position information
and deep belief networks [84], many studies suggested deep of word sequences, Cai et al. [63] and Wen et al. [69] imple-
learning models that capture the spatial-temporal correlation mented the transformer architecture with variations in the
of traffic data. embedding of traffic data and positional encoding. Guo et al.
[64] modified the self-attention score to reflect trends in
A. TEMPORAL FEATURE EXTRACTION traffic data and implemented a dynamic graph convolution
To model time-series traffic data, recurrent neural networks module to replace the position-wise feed-forward layer of the
(RNN) and their variants, such as long-short term memory transformer. TrafficBERT [65] used the transformer encoder
(LSTM) [85] and gated recurrent unit (GRU) [86] have as in Devlin et al. [88] to retain the forecasting power
gained attention in extracting temporal features for traffic when training using data from multiple sources. Wang et al.
forecasting models. Implementation of vanilla LSTM has [72] proposed an approach in which the parameters for the
shown improved performance compared to traditional models self-attention layer is generated using regional distribution
such as auto-regressive integrated moving average (ARIMA), of Point-of-Interests (PoI). Self-attention in conjunction with
other temporal feature extraction methods such as GRU [22], Convolutional GNNs have pioneered GNN-based traffic
[34], [62] and dilated causal convolution [40], [46], [51], [58] forecasting research, and have been widely used in concurrent
have also been proposed. GMAN [68], and AI-GFACN [71] models [15], [16], [17], [22], [28], [29], [31], [32], [33], [35],
adopted self-attention for both spatial and temporal feature [37], [40], [43], [44], [47], [55], [56], [57], [58], [59], [60],
extractions. In addition, Zheng et al. [68] also introduced [62], [63], [64], [67], [71], [75], [95]. Several studies [29],
a transform attention layer that generated spatial-temporal [37], [56], [60] adopted spectral graph convolutions [76], [78]
embedding representations for the positional embedding of that showed higher forecasting power over the basic deep
future time steps. Xu et al. [67] proposed a model in learning models such as feed-forward neural networks and
which a temporal attention block followed a spatial attention FC-LSTM. Li et al. [16] suggested diffusion convolution,
block. The two attention blocks of the model shared similar which expanded the application of graph convolution to
structures, except that the graph convolution operation was directed graphs, and has been applied in many traffic
skip-connected to the output of the spatial attention block to forecasting studies [44], [55], [57], [63]. Cui et al. [15] sug-
reflect the static structure of the transportation network. gested traffic graph convolution (TGC), using element-wise
multiplication between learnable parameters and adjacency
B. SPATIAL FEATURE EXTRACTION WITH GRAPH NEURAL matrices. Zhang et al. [28] implemented traffic graph
NETWORKS convolution with an attention mechanism [96] to capture the
Earlier efforts have adopted CNN to extract spatial features of dependencies in the time steps regardless of distances. Using
traffic data. However, they operate in Euclidean space and fail a matrix factorization technique, Bai et al. [33] suggested
to represent the complex topology of transportation networks a convolutional GNN module that can apply node specific
[20], [23], [25], [26], [39], [42], [48], [49], [50], [54], [89], parameters. Attentional GNNs also have been widely used
[90], [91]. in traffic forecasting research [21], [26], [27], [36], [41],
GNNs have become a popular choice in traffic forecasting [45], [52], [66], [73]. The gated attention networks (GaAN)
since the early adoptions by Li et al. [16] and Yu et al. [26] outperformed diffusion convolution in short-term traffic
[56]. The core idea of GNN is to process the data into forecasting when combined with GRU. GAT [79] has also
graph structures and extract the spatial feature of each node been adopted in many studies [21], [27], [36], [41], [52],
by aggregating the information from neighboring nodes. [73]. Park et al. [66] constructed a new attentional GNN
Most GNN methods for supervised learning, such as classi- layer that adopts the scaled dot-product attention [61] with
fication and regression, can be grouped into convolutional, sentinel vectors to control the information from neighbor
attentional, and message-passing GNNs based on how they nodes. A few studies have implemented convolutional and
aggregate neighborhood information [92]. attentional GNNs in one model [46], [51], [72]. Message-
Convolutional GNNs multiply fixed weights to the source passing GNN traffic forecasting models have also been
node features and conduct aggregating operations, such as suggested using a dual graph that predicts node and edge
summation, pooling, and averaging, to extract target node features [74], and using bidirectional graphs in extracting
spatial features. The most widely used methods under con- aggregated spatial-temporal features [34]. Gupta et al. [38]
volutional GNN are the group of spectral graph convolutions proposed a message-passing GNN-based model with a spatial
[76], [77], [78], which approximates the filters in the spectral embedding and attention mechanism based on shortest-paths
domain. GraphSAGE [93] and diffusion convolution [16] are on graphs. Outside the existing taxonomy of GNNs, graph
other examples of convolutional GNNs. Attentional GNNs embedding techniques such as DeepWalk [97], LINE [98],
resemble convolutional GNNs in that they multiply the source and node2vec [99] have also been adopted to incorporate
node features with scalar weights. The difference, however, graph structures [24], [66], [68], [71], [89].
lies in that the attentional GNNs assign the weights through While these studies have achieved significant performance
a function of the source and target node features. Graph improvements, there have not been sufficient efforts to
attention networks (GAT) [79] and Gated Attention Networks understand the performance of individual building blocks that
(GaAN) [26] are popular attentional GNN models that constitute these models. Li et al. [16] introduced diffusion
implement self-attention mechanisms [61]. Finally, message- convolution as a convolutional GNN layer, employing it
passing GNNs compute output representations of a target to conceptualize traffic flow as a diffusion process occur-
node using a function of the target node and its neighbors. ring on a graph. This approach was then compared with
Gilmer et al. [94] is an example of message-passing GNN, the more traditional ChebNet [76] for their performance.
which computes the message using hidden representations Similarly, Cui et al. [15] conduct a comparative analysis
of source and target nodes and edges. The aggregated between the proposed traffic graph convolution (TGC)
messages and the target node features are passed through a and traditional convolutional GNNs such as spectral GNN
neural network to generate output representations. For more [77] and ChebNet [76] for their number of parameters,
explanations on GNNs taxonomy, see Bronstein et al. [92]. computational efficiency, feature localization ability, and
As transportation networks are inherently equipped with overall performance. For the temporal feature extraction
graph structures, the GNNs have become the most popular blocks, Reza et al. [70] evaluates the performances of the
spatial feature extraction method for traffic forecasting. transformer compared to other machine learning algorithms
A. NOTATIONS AND DEFINITIONS where σ (·) is an activation function, and W ∈ Rd×h is the
Definition 1: Transportation network graph We represent weight parameter matrix where h is the output dimension.
the transportation network graph as a directed graph G = Whereas GCN originally used the normalized Laplacian
−1/2 −1/2
(V , E), where V is a set of |V | = N nodes and E is a matrix  = D̃ ÃD̃ where à = I N + A, and D̃ii =
−1
set of edges representing pairwise connections between the 6j ãij , we use  = D̃ à to apply GCN on directed graphs.
FIGURE 2. Overview of this study. We first define three categories for temporal and two for spatial building blocks.
Combining a spatial and a temporal building block, we implement six models and conduct an extensive and multi-faceted
experiment using four different real-world traffic datasets. Finally, we analyze the results on overall performance,
performance in different traffic categories, performance on outliers, and adaptive model evaluation.
where W z,1 , W z,2 ∈ Rd×d , and bz ∈ Rd are learnable where dij is the distance between sensor vi and vj , σ is the
parameters, and σ (·) is the sigmoid activation. While the standard deviation, and ϵ = 0.1 is the threshold value.
(l) Urban-core and Urban-mix are 5-min speed data for
spatial attention layer to obtain spatial features HS is imple-
mented in a similar manner to the temporal attention layer in road segments in the Seoul traffic network. Both contain
the original work, we replaced the spatial attention module information for one month ranging from April 1, 2018,
with GCN and GAT, denoted GMAN-GCN and GMAN- to April 30, 2018. Urban-core includes 304 records of road
GAT (Fig.6). The transform attention layer is implemented segments in Gangnam, Seoul, one of the regions with the
between the encoder and decoder to enable the multi-step highest traffic and economic activities in the country. The
prediction and reduce error propagation in the prediction task. road segments have similar structural features, such as speed
GMAN can be regarded as a 2-dimensional expansion of the limit, degree, and length.
original transformer [61]. Two parallel self-attention modules Urban-mix is a spatial expansion of Urban-core and
are employed to extract features from both spatial and has road segments with more heterogeneous characteristics.
temporal dimensions, whereas transformer only considers a It contains the inner-city highway connecting the East and
single dimension. To merge representations from two self- West ends of the city, urban arterials, alleys, bridges, and a
attention modules, GMAN replaces the feedforward layer few intercity highway segments. The transportation network
in transformer with a feature fusion layer and makes one graph of Urban-mix has 1,007 road segments. The edges of
residual connection between the input and output of an transportation network graphs are set between road segments
encoder layer. For a more detailed description of GMAN, that share endpoints.
please refer to the original study [68]. When the four datasets are compared in terms of
complexity, the highway flow shows higher complexity
C. DATA than highway speed and urban speed demonstrate higher
To analyze the performance of each model, we select complexity than highway data as in Fig. 7. The approximate
four real-world datasets with diverse characteristics, namely, entropy values [104] on average are 0.52, 1.20, 1.40, and
PeMS-Bay, METR-LA [16],1 Urban-core, and Urban-mix 1.41 for PeMS-Bay, METR-LA, Urban-core, and Urban-mix,
[31].2 respectively. Table 2 summarizes the datasets.
1 PeMS-Bay and METR-LA datasets are available at D. EXPERIMENTAL SETTINGS
https://github.com/liyaguang/DCRNN
2 Urban-core and Urban-mix datasets are available at We adopt mean absolute error (MAE), root mean squared
https://github.com/yuyolshin/SeoulSpeedData error (RMSE), and mean absolute percentage error (MAPE)
FIGURE 7. One day sample data from different datasets. (a) PeMS-Bay, (b) METR-LA, (c) Urban-core, and (d) Urban-mix. The traffic
flow data (METR-LA) shows higher entropy than traffic speed data (PeMS-Bay), and urban data (Urban-core and Urban-mix) show
higher entropy than highway data (PeMS-Bay).
as evaluation metrics for model performances. We calibrated each model hyperparameters as closely as
that in the original works [29], [55], [68]. We set the number
N T′ of hidden units to 64 for GMAN-GCN and GMAN-GAT
1 XX i
MAE = ′ ŷj − yij , (17) and 32 for T-GCN, T-GAT, GWNet-GCN, and GWNet-GAT,
TN batch size to 32, and learning rate to 0.001. For GAT, the
i=1 j=1
v number of heads and dimensions of each head are 8. The
u N T ′ ŷi − yi 2
u
uX X j j number of layers for GMAN models was 3 except for those in
RMSE = t , (18) Urban-mix because of memory limitation and GMAN-GAT
T ′N
i=1 j=1 in METR-LA because the model failed to converge with
N T′ i i 3 layers. A 2-layer model was used in these cases. We trained
1 X X ŷj − yj the models using the Adam optimizer, and L1 loss function.
MAPE = ′ , (19)
TN
i=1 j=1
yij The experiment was conducted on a single NVIDIA TITAN
RTX with 24 GB memory (GPU) and Intel(R) Xeon(R) CPU
ES-2630 v4 @ 2.20 GHz (CPU).3
where T ′ is the total number of predicted time steps, N is the
number of nodes (sensors or road segments), and ŷij and yij are 3 The source codes are available at
the predicted and actual values. https://github.com/yuyolshin/STTFEvaluation
V. RESULTS AND DISCUSSION 70.6%, 63.7%, and 52.4% for the T-GAT, GWNet-GCN, and
A. OVERALL PERFORMANCE GMAN-GAT, respectively. The differences in RMSE in all
Table 3 shows the model performances in the four traffic datasets are presented in Table 4. The self-attention shows
datasets for 15 min (3 steps), 30 min (6 steps), and 60 min robust performance against the increase in prediction horizon,
(12 steps) cases. When combined with the RNN model, yielding a smaller gap between the 15-min and 60-min
GAT-based spatial feature extraction yields more accurate prediction outcomes. This indicates possible advantages for
results than GCN, except for MAE on the 15-min forecast prediction horizons longer than one hour.
in METR-LA. The convolution shows improved predictions
when combined with GCN except for RMSE in Urban-mix B. PERFORMANCE IN DIFFERENT TRAFFIC CATEGORIES
for all prediction horizons. When using self-attention for In this subsection, we analyze the performance of each
temporal feature extraction, GMAN-GAT consistently yields model in different traffic categories. We divided the traffic
improved results than the GCN counterpart on at least states into unequal intervals, considering the range and
one performance metric in all datasets except 15-min and distribution of each dataset. In PeMS-Bay, we initially
30-min predictions in METR-LA. Overall, the convolution divided the speed data with equal intervals of 10 mph.
models yield the best performance among the comparative However, we merged the five lower speed intervals because
models except in long-term (60-min) prediction in PeMS-Bay each interval contained few observations, and merged the two
and Urban-core. Although T-GAT produces fair prediction higher speed intervals for the same reason. Since the 60∼70
outcomes, RNN shows no clear advantage over the other mph interval included nearly 80% of the data, we divided
building blocks for temporal feature extraction. the interval into two intervals of 5 mph. Finally, we have
The three temporal building blocks methods show dif- five speed categories in PeMS-Bay: 0∼50 mph, 50∼60 mph,
ferences in the gap between the forecasting accuracy on 60∼65 mph, 65∼70 mph, and 70∼90 mph.
the 15-min and 60-min predictions. The RMSE differences Table 5 presents the results of the traffic forecasting models
between the two prediction horizons in PeMS-Bay are in PeMS-Bay, across different traffic speed categories and
TABLE 4. RMSE gap between the 15-min and 60-min prediction for all datasets. The attention-based GMAN models show smaller gaps compared to the
other models.
prediction horizons. The best performance is observed in the TABLE 5. Performance in MAE by traffic speed categories in PeMS-Bay.
GWNet-GCN presents high performances in high-frequency categories,
65∼70 mph category, which contains the most observations. while GMAN-GAT performs better in low-frequency categories.
In contrast, the largest errors are observed in the 0∼50
mph category, which is furthest from the high-speed, high-
frequency 65∼70 mph category. In categories with over
60 mph, the category-wise errors are smaller than the overall
performance. Similar to the overall performance evaluation,
the two models outperform the RNN model. The convolution-
based GWNet-GCN achieved high performances in the high-
frequency categories. For 60-min prediction, GWNet-GCN
produces more accurate predictions than GMAN-GAT in
60∼65, 65∼70, and 70∼90 mph categories. In contrast,
GMAN-GAT shows more robust performance across dif-
ferent traffic categories than GWNet-GCN. In PeMS-Bay,
the 0∼50 mph category MAE is 9.9 times larger than the
65∼70 mph category MAE for GWNet-GCN on 60-min
prediction. In contrast, the ratio is 7.9 for GMAN-GAT. The
TABLE 6. Prediction results of 60-min during traffic transitions.
ratios are 6.9 and 6.4 on 15-min predictions for GWNet-GCN
and GMAN-GAT, respectively. Similar trends are observed
in other datasets. In METR-LA, GWNet-GCN performs
better in high-frequency categories (60∼65 and 65∼75
veh/h), while GMAN-GAT shows higher performance in
low-frequency categories (30∼50 and 50∼60 veh/h). In the
0∼30 veh/h category, the self-attention model performance
decreases, and the convolution model performance improves.
For Urban-core, the distributions are right-skewed as opposed
to highway datasets. Therefore, convolution models are more
effective at low-speed categories, whereas self-attention mod-
els are better suited for high-speed categories. In Urban-mix,
GWNet-GCN achieves the highest performance across all
speed categories and prediction horizons. The category-wise
performances for METR-LA, Urban-core, and Urban-mix are
presented in Fig. 8.
We also analyze model performances in conditions where
traffic states experience transitions. We denote the condition
where the speed increases or decreases more than 30 mph the performance of 60-min forecasting outcomes during
in 90 min (18 time steps) in PeMS-Bay as speed increase traffic transitions for all datasets. The results on the other
and decrease transitions, respectively, and compare the datasets show similar trends as for PeMS-Bay. The RNN and
60-min prediction results. During transitions, the model self-attention models show advantages over the convolution
performances differ from the overall performances. Whereas model except in the Urban-mix. The traffic transition
GWNet-GCN yielded low MAE and MAPE overall, GMAN- conditions in the other datasets are defined if the states
GAT outperformed GWNet-GCN in all performance metrics change by 30 veh/h, 10 km/h, and 20 km/h for METR-LA,
in increasing and decreasing transitions. Table 6 presents Urban-core, and Urban-mix, respectively.
FIGURE 8. Performance by traffic flow and speed categories on (a) METR-LA, (b) Urban-core, and (c) Urban-mix. The line
graphs are the MAE of each model in each traffic category, and the red histogram in the background is the ratio of each
category in each dataset. Among the three models, GWNet-GCN achieves the best performance in categories with high
observation percentage, and GMAN-GAT generally achieves the best performance in categories with low observation
percentage.
C. ROBUSTNESS AGAINST OUTLIERS each prediction horizon l by averaging the two candidate
Another characteristic observed is robustness against outliers models, G1 and G2 , as follows:
in the labels. As in Figs. 9(a), (d), 10 (a) and (d), RNN
and convolution-based temporal feature extractions show l 1
Ŷ (p) = G1 (X )l + G2 (X )l , (20)
delayed reactions to outliers, causing large errors within 2
a few time steps. While GWNet-GCN shows the highest
overall accuracy in most datasets and prediction horizons where Xval is the input data of validation sets, and (G(Xval )l
as presented in previous sections, the self-attention model is the output of the model G for prediction horizon l. The
is more robust against outliers than the other models. pseudo-labels are necessary to distribute the test sets in which
In addition, the attention-based GAT models also show more category they should be evaluated. For each traffic category s,
robustness than GCN models for spatial feature extraction, we compare the loss for the two models and make predictions
as shown in Fig. 11. Ŷtest l,s as follows:
D. ADAPTIVE MODEL EVALUATION
α ∗ G1 (Xtest
s l
) + (1 − α) ∗ G2 (Xtest
s l
)
if L(G (X s )l , Y (p)l,s ) > L(G (X s )l , Y (p)l,s )
The model performance was found to change by traffic state
1 val val 2 val val
categories. In this section, we evaluate the models adaptively Ŷtest l,s =
(1 − α) ∗ G (X
1 test
s l
) + α ∗ G (X s l
2 test )
by selecting the model depending on the prediction horizon
(p)l,s (p)l,s
s l
) , Yval ) < L(G2 (Xval
s l
) , Yval )
if L(G1 (Xval
and traffic category-wise performance on validation sets. For
l
adaptive evaluation, we first make pseudo-labels Ŷ (p) for (21)
FIGURE 9. 60-min prediction labels (blue) and outcomes of T-GAT (orange), GWNet-GCN (green), and GMAN-GCN (red)
in METR-LA for sample nodes. In (a) and (d), the RNN and convolution models show delayed reactions to outliers.
While the convolution model shows the highest overall accuracy for 60-min prediction in METR-LA, the attention
model shows more robustness against outliers.
FIGURE 10. 60-min prediction labels (blue) and outcomes of T-GAT (orange), GWNet-GCN (green), and GMAN-GAT
(red) in Urban-core for sample nodes. In (a) and (d), the RNN and convolution models show delayed reactions to
outliers. In Urban-core, the attention model achieves more robustness against outliers compared to the other models
along with the highest accuracy for 60-min prediction.
l,s
where α is a predefined value between 0.5 and 1, Y (p) is performance metrics in all datasets and prediction horizons
the pseudo-label for prediction horizon l included in category as shown in Table 7. For 60-min prediction in PeMS-Bay, the
l,s
s, and X s is the corresponding pseudo-label Y (p) . For the performance gain is the largest, outperforming the previous
final prediction, we calculate Eq. (21) for all categories and state-of-the-art GMAN and GWNet by 3.7%. When the
aggregate the category-wise results. In this experiment, α is Diebold-Mariano test is conducted for 60-min forecasts,
set to 0.7. The concept of this adaptive model evaluation forecasts on 57.5%, 44.4%, 31.3%, and 56.8% of nodes are
framework is visualized in Fig. 12. The adaptive evaluation statistically significant (α = 0.1) in PeMS-Bay, METR-LA,
framework achieved higher performance on at least two Urban-core, and Urban-mix, respectively.
FIGURE 11. Robustness against outliers by different spatial feature extraction methods. GAT models show more
robustness against GCN models for all T-GCN, GWNet, and GMAN models.
FIGURE 12. Adaptive model evaluation framework. It adaptively selects which model to conduct prediction for different
traffic categories based on the performance on validations sets. As a result, the prediction can be made with multiple
models, improving the utility of each model.
We infer that these paired methods are similar in extracting Sophisticated state-of-the-art models could be investigated
information from input data. to discover whether the model characteristics would persist.
Further assessments reveal that the models show differ- Explainable artificial intelligence techniques [105], [106]
ent performance sensitivity to traffic state changes. The could also be adopted to explore the deep learning-based
convolution model performed well in high-frequency traffic traffic forecasting model characteristics. These techniques
categories, and the self-attention model showed robust have been rarely used in traffic forecasting studies [107]
performances even in low-frequency traffic categories and and could give a new direction if implemented appropriately.
with outliers. In addition, during the traffic transitions, the Moreover, the adaptive model evaluation framework will be
self-attention, and RNN models show advantages in long- refined to include predictions during transition states and
term prediction. The attention-based methods in spatial against time-series anomalies.
and temporal dimensions demonstrated improved robustness
with outliers. Overall, the convolution model achieves more GENERATIVE AI AND AI-ASSISTED TECHNOLOGIES
performance gain for the short-term (15-min) prediction During the preparation of this work the authors used ChatGPT
and high-frequency traffic categories. In contrast, the and Grammarly in order to check the grammar. After using
self-attention model has more advantages in prediction for this tool/service, the authors reviewed and edited the content
less-informed conditions such as longer prediction horizons, as needed and take full responsibility for the content of the
low-frequency traffic categories, and outliers. publication.
In addition, we suggest a framework that adaptively selects
a model for each category to make predictions based on the
CONFLICT OF INTEREST
validation set performance. The results reveal that the simple
The authors declare that there is no conflict of interest
implementation of an adaptive evaluation framework could
regarding the publication of this paper.
improve the performance of the previous state-of-the-art by
3.7% at most. This framework enhances traffic forecasting
performance using the existing models rather than developing REFERENCES
more sophisticated models. [1] E. I. Vlahogianni, M. G. Karlaftis, and J. C. Golias, ‘‘Short-term traffic
forecasting: Where we are and where we’re going,’’ Transp. Res. C,
Emerg. Technol., vol. 43, pp. 3–19, Jun. 2014.
VI. CONCLUSION [2] L. Zhu, F. R. Yu, Y. Wang, B. Ning, and T. Tang, ‘‘Big data analytics in
In this study, we investigated the characteristics and evaluated intelligent transportation systems: A survey,’’ IEEE Trans. Intell. Transp.
Syst., vol. 20, no. 1, pp. 383–398, Jan. 2019.
the performance of building blocks of spatial-temporal deep
[3] K. Lee, M. Eo, E. Jung, Y. Yoon, and W. Rhee, ‘‘Short-term traffic
learning models for traffic forecasting. We implemented six prediction with deep neural networks: A survey,’’ IEEE Access, vol. 9,
spatial-temporal models using two spatial building blocks and pp. 54739–54756, 2021.
three temporal building blocks and conducted a multi-faceted [4] A. Sumalee, P. Luathep, W. H. K. Lam, and R. D. Connors, ‘‘Evaluation
and design of transport network capacity under demand uncertainty,’’
experiment analyzing the overall performance, category- Transp. Res. Rec., J. Transp. Res. Board, vol. 2090, no. 1, pp. 17–28,
wise performance, and robustness against outliers. The Jan. 2009.
models were tested on four real-world datasets with diverse [5] X. Fang, J. Huang, F. Wang, L. Zeng, H. Liang, and H. Wang,
transportation networks. While GWNet-GCN demonstrated ‘‘ConSTGAT: Contextual spatial–temporal graph attention network for
travel time estimation at Baidu Maps,’’ in Proc. 26th ACM SIGKDD Int.
the most accurate overall performance in most datasets Conf. Knowl. Discovery Data Mining, Aug. 2020, pp. 2697–2705.
and prediction horizons, GMAN-GAT showed a similar [6] C.-Y. Jiang, X.-M. Hu, and W.-N. Chen, ‘‘An urban traffic signal control
performance level with GWNet-GCN for 60-min prediction system based on traffic flow prediction,’’ in Proc. 13th Int. Conf. Adv.
Comput. Intell. (ICACI), May 2021, pp. 259–265.
in PeMS-Bay and outperformed GWNet-GCN for 60-min [7] D. Rolnick et al., ‘‘Tackling climate change with machine learning,’’ ACM
prediction in Urban-core. Further investigations revealed Comput. Surv., vol. 55, no. 2, pp. 1–96, Feb. 2022.
that the self-attention model had stronger robustness against [8] M. Jusup, P. Holme, K. Kanazawa, M. Takayasu, I. Romic, Z. Wang,
data imbalance and outliers than the RNN and convolution S. Gecek, T. Lipic, B. Podobnik, L. Wang, W. Luo, T. Klanjscek, J. Fan,
S. Boccaletti, and M. Perc, ‘‘Social physics,’’ Phys. Rep., vol. 948,
models. GAT models showed more robustness than GCN pp. 1–148, Feb. 2022.
models amongst spatial feature extraction methods. Finally, [9] B. M. Williams and L. A. Hoel, ‘‘Modeling and forecasting vehicular
an adaptive model evaluation framework demonstrated the traffic flow as a seasonal ARIMA process: Theoretical basis and empirical
results,’’ J. Transp. Eng., vol. 129, no. 6, pp. 664–672, Nov. 2003.
enhanced performance of the existing models without
[10] S. R. Chandra and H. Al-Deek, ‘‘Predictions of freeway traffic speeds
sophistication in model architecture. and volumes using vector autoregressive models,’’ J. Intell. Transp. Syst.,
In the future study, we aim to expand the scope of this vol. 13, no. 2, pp. 53–72, May 2009.
comparative study on building blocks to include sensitivity [11] C.-H. Wu, J.-M. Ho, and D. T. Lee, ‘‘Travel-time prediction with support
vector regression,’’ IEEE Trans. Intell. Transp. Syst., vol. 5, no. 4,
analysis of the hyperparameters and impact analysis of pp. 276–281, Dec. 2004.
different input features such as daily and weekly trends, [12] G. A. Davis and N. L. Nihan, ‘‘Nonparametric regression and short-term
and multi-channel inputs. Also, revealing characteristics of freeway traffic forecasting,’’ J. Transp. Eng., vol. 117, no. 2, pp. 178–188,
Mar. 1991.
additional spatial feature extraction methods such as diffusion
[13] R. Chrobok, J. Wahle, and M. Schreckenberg, ‘‘Traffic forecast using
convolution [16], traffic graph convolution [15], and adaptive simulations of large scale networks,’’ in Proc. IEEE Intell. Transp. Syst.
graph convolution [44], [55], [57] can be another objective. (ITSC), Aug. 2001, pp. 434–439.
[14] S. Jeon and B. Hong, ‘‘Monte Carlo simulation-based traffic speed [34] Y. Wang, J. Zheng, Y. Du, C. Huang, and P. Li, ‘‘Traffic-GGNN:
forecasting using historical big data,’’ Future Gener. Comput. Syst., Predicting traffic flow via attentional spatial–temporal gated graph
vol. 65, pp. 182–195, Dec. 2016. neural networks,’’ IEEE Trans. Intell. Transp. Syst., vol. 23, no. 10,
[15] Z. Cui, K. Henrickson, R. Ke, and Y. Wang, ‘‘Traffic graph convolutional pp. 18423–18432, Oct. 2022.
recurrent neural network: A deep learning framework for network-scale [35] Y. Chen, I. Segovia, and Y. R. Gel, ‘‘Z-GCNETs: Time zigzags at graph
traffic learning and forecasting,’’ IEEE Trans. Intell. Transp. Syst., vol. 21, convolutional networks for time series forecasting,’’ in Proc. 38th Int.
no. 11, pp. 4883–4894, Nov. 2020. Conf. Mach. Learn., Jul. 2021, pp. 1684–1694.
[16] Y. Li, R. Yu, C. Shahabi, and Y. Liu, ‘‘Diffusion convolutional recurrent [36] H. Dong, P. Zhu, J. Gao, L. Jia, and Y. Qin, ‘‘A short-term traffic flow
neural network: Data-driven traffic forecasting,’’ in Proc. 6th Int. Conf. forecasting model based on spatial–temporal attention neural network,’’
Learn. Represent., May 2018, pp. 1–16. in Proc. IEEE 25th Int. Conf. Intell. Transp. Syst. (ITSC), Oct. 2022,
[17] B. Liao, J. Zhang, C. Wu, D. McIlwraith, T. Chen, S. Yang, Y. Guo, and pp. 416–421.
F. Wu, ‘‘Deep sequence learning with auxiliary information for traffic [37] Q. Zhang, M. Tan, C. Li, H. Xia, W. Chang, and M. Li, ‘‘Spatio-
prediction,’’ in Proc. 24th ACM SIGKDD Int. Conf. Knowl. Discovery temporal residual graph convolutional network for short-term traffic flow
Data Mining, Jul. 2018, pp. 537–546. prediction,’’ IEEE Access, vol. 11, pp. 84187–84199, 2023.
[18] X. Ma, Z. Tao, Y. Wang, H. Yu, and Y. Wang, ‘‘Long short-term memory [38] M. Gupta, H. Kodamana, and S. Ranu, ‘‘Frigate: Frugal spatio-temporal
neural network for traffic speed prediction using remote microwave forecasting on road networks,’’ in Proc. 29th ACM SIGKDD Conf. Knowl.
sensor data,’’ Transp. Res. C, Emerg. Technol., vol. 54, pp. 187–197, Discovery Data Mining, Aug. 2023, pp. 649–660.
May 2015. [39] C. Chen, K. Li, S. G. Teo, G. Chen, X. Zou, X. Yang, R. C. Vijay, J. Feng,
[19] X. Ma, H. Yu, Y. Wang, and Y. Wang, ‘‘Large-scale transportation and Z. Zeng, ‘‘Exploiting spatio-temporal correlations with multiple 3D
network congestion evolution prediction using deep learning theory,’’ convolutional neural networks for citywide vehicle flow prediction,’’ in
PLoS ONE, vol. 10, no. 3, Mar. 2015, Art. no. e0119044. Proc. IEEE Int. Conf. Data Mining (ICDM), Nov. 2018, pp. 893–898.
[20] X. Ma, J. Zhang, B. Du, C. Ding, and L. Sun, ‘‘Parallel architecture [40] S. Fang, Q. Zhang, G. Meng, S. Xiang, and C. Pan, ‘‘GSTNet: Global
of convolutional bi-directional LSTM neural networks for network-wide spatial–temporal network for traffic flow prediction,’’ in Proc. 28th Int.
metro ridership prediction,’’ IEEE Trans. Intell. Transp. Syst., vol. 20, Joint Conf. Artif. Intell., Aug. 2019, pp. 2286–2293.
no. 6, pp. 2278–2288, Jun. 2019. [41] Y. Fang, Y. Qin, H. Luo, F. Zhao, B. Xu, C. Wang, and L. Zeng,
[21] Z. Pan, Y. Liang, W. Wang, Y. Yu, Y. Zheng, and J. Zhang, ‘‘Urban traffic ‘‘Spatio-temporal meets wavelet: Disentangled traffic flow forecasting
prediction from spatio-temporal data using deep meta learning,’’ in Proc. via efficient spectral graph attention network,’’ 2021, arXiv:2112.02740.
25th ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, Jul. 2019, [42] S. Guo, Y. Lin, S. Li, Z. Chen, and H. Wan, ‘‘Deep spatial–temporal 3D
pp. 1720–1730. convolutional neural networks for traffic data forecasting,’’ IEEE Trans.
Intell. Transp. Syst., vol. 20, no. 10, pp. 3913–3926, Oct. 2019.
[22] X. Wang, Y. Ma, Y. Wang, W. Jin, X. Wang, J. Tang, C. Jia, and J. Yu,
‘‘Traffic flow prediction via spatial temporal graph neural network,’’ [43] L. Ge, S. Li, Y. Wang, F. Chang, and K. Wu, ‘‘Global spatial–temporal
in Proc. Web Conf. New York, NY, USA: Association for Computing graph convolutional network for urban traffic speed prediction,’’ Appl.
Machinery, Apr. 2020, pp. 1082–1092. Sci., vol. 10, no. 4, p. 1509, Feb. 2020.
[44] L. Han, B. Du, L. Sun, Y. Fu, Y. Lv, and H. Xiong, ‘‘Dynamic and multi-
[23] Y. Wu, H. Tan, L. Qin, B. Ran, and Z. Jiang, ‘‘A hybrid
faceted spatio-temporal deep learning for traffic speed forecasting,’’
deep learning based traffic flow prediction method and its
in Proc. 27th ACM SIGKDD Conf. Knowl. Discovery Data Mining,
understanding,’’ Transp. Res. C, Emerg. Technol., vol. 90, pp. 166–180,
Aug. 2021, pp. 547–555.
May 2018.
[45] R. Huang, C. Huang, Y. Liu, G. Dai, and W. Kong, ‘‘LSGCN: Long short-
[24] D. Xu, H. Dai, Y. Wang, P. Peng, Q. Xuan, and H. Guo, ‘‘Road traffic
term traffic prediction with graph convolutional networks,’’ in Proc. 29th
state prediction based on a graph embedding recurrent neural network
Int. Joint Conf. Artif. Intell., Jul. 2020, pp. 2355–2361.
under the SCATS,’’ Chaos, Interdiscipl. J. Nonlinear Sci., vol. 29, no. 10,
[46] B. Lu, X. Gan, H. Jin, L. Fu, and H. Zhang, ‘‘Spatiotemporal adaptive
Oct. 2019, Art. no. 103125.
gated graph convolution network for urban traffic flow forecasting,’’
[25] H. Yu, Z. Wu, S. Wang, Y. Wang, and X. Ma, ‘‘Spatiotemporal recurrent
in Proc. 29th ACM Int. Conf. Inf. Knowl. Manage., Oct. 2020,
convolutional networks for traffic prediction in transportation networks,’’
pp. 1025–1034.
Sensors, vol. 17, no. 7, p. 1501, Jun. 2017.
[47] K. Lee and W. Rhee, ‘‘DDP-GCN: Multi-graph convolutional network
[26] J. Zhang, X. Shi, J. Xie, H. Ma, I. King, and D. Y. Yeung, ‘‘GaAN: for spatiotemporal traffic forecasting,’’ Transp. Res. C, Emerg. Technol.,
Gated attention networks for learning on large and spatiotemporal vol. 134, Jan. 2022, Art. no. 103466.
graphs,’’ in Proc. 34th Conf. Uncertain Artif. Intell., Aug. 2018,
[48] Q. Liu, B. Wang, and Y. Zhu, ‘‘Short-term traffic speed forecasting based
2018.
on attention convolutional neural network for arterials,’’ Comput.-Aided
[27] C. Zhang, J. J. Q. Yu, and Y. Liu, ‘‘Spatial–temporal graph attention Civil Infrastruct. Eng., vol. 33, no. 11, pp. 999–1016, Nov. 2018.
networks: A deep learning approach for traffic forecasting,’’ IEEE Access, [49] X. Ma, Z. Dai, Z. He, J. Ma, Y. Wang, and Y. Wang, ‘‘Learning
vol. 7, pp. 166246–166256, 2019. traffic as images: A deep convolutional neural network for large-scale
[28] Z. Zhang, M. Li, X. Lin, Y. Wang, and F. He, ‘‘Multistep speed prediction transportation network speed prediction,’’ Sensors, vol. 17, no. 4, p. 818,
on traffic networks: A deep learning approach considering spatio- Apr. 2017.
temporal dependencies,’’ Transp. Res. C, Emerg. Technol., vol. 105, [50] G. Shen, C. Chen, Q. Pan, S. Shen, and Z. Liu, ‘‘Research on traffic
pp. 297–322, Aug. 2019. speed prediction by temporal clustering analysis and convolutional neural
[29] L. Zhao, Y. Song, C. Zhang, Y. Liu, P. Wang, T. Lin, M. Deng, network with deformable kernels (May, 2018),’’ IEEE Access, vol. 6,
and H. Li, ‘‘T-GCN: A temporal graph convolutional network for pp. 51756–51765, 2018.
traffic prediction,’’ IEEE Trans. Intell. Transp. Syst., vol. 21, no. 9, [51] Q. Song, R. Ming, J. Hu, H. Niu, and M. Gao, ‘‘Graph attention convolu-
pp. 3848–3858, Sep. 2020. tional network: Spatiotemporal modeling for urban traffic prediction,’’
[30] Z. Zhao, W. Chen, X. Wu, P. C. Y. Chen, and J. Liu, ‘‘LSTM network: in Proc. IEEE 23rd Int. Conf. Intell. Transp. Syst. (ITSC), Sep. 2020,
A deep learning approach for short-term traffic forecast,’’ IET Intell. pp. 1–6.
Transp. Syst., vol. 11, no. 2, pp. 68–75, Mar. 2017. [52] K. Tian, J. Guo, K. Ye, and C.-Z. Xu, ‘‘ST-MGAT: Spatial–temporal
[31] Y. Shin and Y. Yoon, ‘‘Incorporating dynamicity of transportation multi-head graph attention networks for traffic forecasting,’’ in Proc.
network with multi-weight traffic graph convolutional network for IEEE 32nd Int. Conf. Tools with Artif. Intell. (ICTAI), Nov. 2020,
traffic forecasting,’’ IEEE Trans. Intell. Transp. Syst., vol. 23, no. 3, pp. 714–721.
pp. 2082–2092, Mar. 2022. [53] C. Tian and W. K. V. Chan, ‘‘Spatial–temporal attention Wavenet: A deep
[32] X. Huang, Y. Ye, X. Yang, and L. Xiong, ‘‘Multi-view dynamic graph learning framework for traffic prediction considering spatial–temporal
convolution neural network for traffic flow prediction,’’ Expert Syst. dependencies,’’ IET Intell. Transp. Syst., vol. 15, no. 4, pp. 549–561,
Appl., vol. 222, Jul. 2023, Art. no. 119779. Apr. 2021.
[33] L. Bai, L. Yao, C. Li, X. Wang, and C. Wang, ‘‘Adaptive graph [54] J. Wang, Q. Gu, J. Wu, G. Liu, and Z. Xiong, ‘‘Traffic speed prediction
convolutional recurrent network for traffic forecasting,’’ in Proc. Adv. and congestion source exploration: A deep learning method,’’ in Proc.
Neural Inf. Process. Syst., vol. 33, Dec. 2020, pp. 17804–17815. IEEE 16th Int. Conf. Data Mining (ICDM), Dec. 2016, pp. 499–508.
[55] Z. Wu, S. Pan, G. Long, J. Jiang, and C. Zhang, ‘‘Graph WaveNet for deep [78] T. Kipf and M. Welling, ‘‘Semi-supervised classification with graph
spatial–temporal graph modeling,’’ in Proc. 28th Int. Joint Conf. Artif. convolutional networks,’’ in Proc. 5th Int. Conf. Learn. Represent.,
Intell., Aug. 2019, pp. 1907–1913. Apr. 2017, pp. 1–14.
[56] B. Yu, H. Yin, and Z. Zhu, ‘‘Spatio-temporal graph convolutional [79] P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Lio, and
networks: A deep learning framework for traffic forecasting,’’ in Proc. Y. Bengio, ‘‘Graph attention networks,’’ in Proc. 6th Int. Conf. Learn.
27th Int. Joint Conf. Artif. Intell., Jul. 2018, pp. 3634–3640. Represent., May 2018, pp. 1–12.
[57] Y. Shin and Y. Yoon, ‘‘PGCN: Progressive graph convolutional networks [80] A. Krizhevsky, I. Sutskever, and G. E. Hinton, ‘‘ImageNet classification
for spatial–temporal traffic forecasting,’’ 2022, arXiv:2202.08982. with deep convolutional neural networks,’’ in Proc. Adv. Neural Inf.
[58] J. Zhao, Z. Liu, Q. Sun, Q. Li, X. Jia, and R. Zhang, ‘‘Attention-based Process. Syst., vol. 25, Dec. 2012, pp. 1–9.
dynamic spatial–temporal graph convolutional networks for traffic speed [81] R. Girshick, J. Donahue, T. Darrell, and J. Malik, ‘‘Rich feature
forecasting,’’ Expert Syst. Appl., vol. 204, Oct. 2022, Art. no. 117511. hierarchies for accurate object detection and semantic segmentation,’’ in
[59] K. Zhu, S. Zhang, J. Li, D. Zhou, H. Dai, and Z. Hu, ‘‘Spatiotemporal Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2014, pp. 580–587.
multi-graph convolutional networks with synthetic data for traffic volume [82] I. Sutskever, O. Vinyals, and Q. V. Le, ‘‘Sequence to sequence learning
forecasting,’’ Expert Syst. Appl., vol. 187, Jan. 2022, Art. no. 115992. with neural networks,’’ in Advances in Neural Information Processing
[60] B. Chen, K. Hu, Y. Li, and L. Miao, ‘‘Hybrid spatio-temporal graph Systems, vol. 27, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence,
convolution network for short-term traffic forecasting,’’ in Proc. IEEE and K. Weinberger, Eds. Red Hook, NY, USA: Curran Associates, 2014.
25th Int. Conf. Intell. Transp. Syst. (ITSC), Oct. 2022, pp. 2128–2133. [83] Y. Lv, Y. Duan, W. Kang, Z. Li, and F.-Y. Wang, ‘‘Traffic flow prediction
[61] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, with big data: A deep learning approach,’’ IEEE Trans. Intell. Transp.
L. U. Kaiser, and I. Polosukhin, ‘‘Attention is all you need,’’ in Proc. Adv. Syst., vol. 16, no. 2, pp. 865–873, Apr. 2015.
Neural Inf. Process. Syst., vol. 30, Dec. 2017, pp. 1–11. [84] W. Huang, G. Song, H. Hong, and K. Xie, ‘‘Deep architecture for traffic
flow prediction: Deep belief networks with multitask learning,’’ IEEE
[62] J. Bai, J. Zhu, Y. Song, L. Zhao, Z. Hou, R. Du, and H. Li, ‘‘A3T-GCN:
Trans. Intell. Transp. Syst., vol. 15, no. 5, pp. 2191–2201, Oct. 2014.
Attention temporal graph convolutional network for traffic forecasting,’’
ISPRS Int. J. Geo-Inf., vol. 10, no. 7, p. 485, Jul. 2021. [85] S. Hochreiter and J. Schmidhuber, ‘‘Long short-term memory,’’ Neural
Comput., vol. 9, no. 8, pp. 1735–1780, Nov. 1997.
[63] L. Cai, K. Janowicz, G. Mai, B. Yan, and R. Zhu, ‘‘Traffic transformer:
Capturing the continuity and periodicity of time series for traffic [86] K. Cho, B. van Merrienboer, D. Bahdanau, and Y. Bengio, ‘‘On the
forecasting,’’ Trans. GIS, vol. 24, no. 3, pp. 736–755, Jun. 2020. properties of neural machine translation: Encoder–decoder approaches,’’
2014, arXiv:1409.1259.
[64] S. Guo, Y. Lin, H. Wan, X. Li, and G. Cong, ‘‘Learning dynamics and
[87] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals,
heterogeneity of spatial–temporal graph data for traffic forecasting,’’
A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, ‘‘WaveNet:
IEEE Trans. Knowl. Data Eng., vol. 34, no. 11, pp. 5415–5428,
A generative model for raw audio,’’ 2016, arXiv:1609.03499.
Nov. 2022.
[88] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, ‘‘BERT: Pre-training
[65] K. Jin, J. Wi, E. Lee, S. Kang, S. Kim, and Y. Kim, ‘‘TrafficBERT:
of deep bidirectional transformers for language understanding,’’ 2018,
Pre-trained model with large-scale data for long-range traffic flow
arXiv:1810.04805.
forecasting,’’ Expert Syst. Appl., vol. 186, Dec. 2021, Art. no. 115738.
[89] H. Yao, F. Wu, J. Ke, X. Tang, Y. Jia, S. Lu, P. Gong, J. Ye, and
[66] C. Park, C. Lee, H. Bahng, Y. Tae, S. Jin, K. Kim, S. Ko, and J. Choo, ‘‘ST-
Z. Li, ‘‘Deep multi-view spatial–temporal network for taxi demand
GRAT: A novel spatio-temporal graph attention networks for accurately
prediction,’’ in Proc. 32nd AAAI Conf. Artif. Intell., 2018, vol. 32, no. 1,
forecasting dynamically changing road speed,’’ in Proc. 29th ACM Int.
pp. 2588–2595.
Conf. Inf. Knowl. Manage., Oct. 2020, pp. 1215–1224.
[90] H. Yao, X. Tang, H. Wei, G. Zheng, and Z. Li, ‘‘Revisiting spatial–
[67] M. Xu, W. Dai, C. Liu, X. Gao, W. Lin, G.-J. Qi, and H. Xiong, temporal similarity: A deep learning framework for traffic prediction,’’ in
‘‘Spatial–temporal transformer networks for traffic flow forecasting,’’ Proc. 33rd AAAI Conf. Artif. Intell., 2019, vol. 33, no. 1, pp. 5668–5675.
2020, arXiv:2001.02908.
[91] Z. Lin, J. Feng, Z. Lu, Y. Li, and D. Jin, ‘‘DeepSTN+: Context-
[68] C. Zheng, X. Fan, C. Wang, and J. Qi, ‘‘GMAN: A graph multi-attention aware spatial–temporal neural network for crowd flow prediction in
network for traffic prediction,’’ in Proc. 34th AAAI Conf. Artif. Intell., metropolis,’’ in Proc. 33rd AAAI conf. Artif. intell., 2019, vol. 33, no. 1,
Feb. 2020, vol. 34, no. 1, pp. 1234–1241. pp. 1020–1027.
[69] Y. Wen, P. Xu, Z. Li, W. Xu, and X. Wang, ‘‘RPConvformer: A novel [92] M. M. Bronstein, J. Bruna, T. Cohen, and P. Veličković, ‘‘Geometric
transformer-based deep neural networks for traffic flow prediction,’’ deep learning: Grids, groups, graphs, geodesics, and gauges,’’ 2021,
Expert Syst. Appl., vol. 218, May 2023, Art. no. 119587. arXiv:2104.13478.
[70] S. Reza, M. C. Ferreira, J. J. M. Machado, and J. M. R. S. Tavares, [93] W. Hamilton, Z. Ying, and J. Leskovec, ‘‘Inductive representation
‘‘A multi-head attention-based transformer model for traffic flow learning on large graphs,’’ in Proc. Adv. Neural Inf. Process. Syst., vol. 30,
forecasting with a comparative analysis to recurrent neural networks,’’ 2017, pp. 1–11.
Expert Syst. Appl., vol. 202, Sep. 2022, Art. no. 117275. [94] J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl,
[71] X. Xu, X. Hu, Y. Zhao, X. Lü, and A. Aapaoja, ‘‘Urban short-term traffic ‘‘Neural message passing for quantum chemistry,’’ in Proc. 34th Int. Conf.
speed prediction with complicated information fusion on accidents,’’ Mach. Learn., Aug. 2017, pp. 1263–1272.
Expert Syst. Appl., vol. 224, Aug. 2023, Art. no. 119887. [95] C. Song, Y. Lin, S. Guo, and H. Wan, ‘‘Spatial–temporal synchronous
[72] K. Wang, L. Liu, Y. Liu, G. Li, F. Zhou, and L. Lin, ‘‘Urban regional graph convolutional networks: A new framework for spatial–temporal
function guided traffic flow prediction,’’ Inf. Sci., vol. 634, pp. 308–320, network data forecasting,’’ in Proc. 34th AAAI Conf. Artif. Intell., 2020,
Jul. 2023. vol. 34, no. 1, pp. 914–921.
[73] X. Kong, W. Xing, X. Wei, P. Bao, J. Zhang, and W. Lu, ‘‘STGAT: [96] D. Bahdanau, K. Cho, and Y. Bengio, ‘‘Neural machine translation by
Spatial–temporal graph attention networks for traffic flow forecasting,’’ jointly learning to align and translate,’’ in Proc. 3rd Int. Conf. Learn.
IEEE Access, vol. 8, pp. 134363–134372, 2020. Represent., 2015, pp. 1–15.
[74] L. Wei, Z. Yu, Z. Jin, L. Xie, J. Huang, D. Cai, X. He, and [97] B. Perozzi, R. Al-Rfou, and S. Skiena, ‘‘DeepWalk: Online learning of
X.-S. Hua, ‘‘Dual graph for traffic forecasting,’’ IEEE Access, early social representations,’’ in Proc. 20th ACM SIGKDD Int. Conf. Knowl.
access, Dec. 9, 2019, doi: 10.1109/ACCESS.2019.2958380. Discovery Data Mining, Aug. 2014, pp. 701–710.
[75] S. Li, L. Ge, Y. Lin, and B. Zeng, ‘‘Adaptive spatial–temporal fusion graph [98] J. Tang, M. Qu, M. Wang, M. Zhang, J. Yan, and Q. Mei, ‘‘LINE: Large-
convolutional networks for traffic flow forecasting,’’ in Proc. Int. Joint scale information network embedding,’’ in Proc. 24th Int. Conf. World
Conf. Neural Netw. (IJCNN), Jul. 2022, pp. 4189–4196. Wide Web, May 2015, pp. 1067–1077.
[76] M. Defferrard, X. Bresson, and P. Vandergheynst, ‘‘Convolutional neural [99] A. Grover and J. Leskovec, ‘‘node2vec: Scalable feature learning for
networks on graphs with fast localized spectral filtering,’’ in Proc. Adv. networks,’’ in Proc. 22nd ACM SIGKDD Int. Conf. Knowl. Discovery
Neural Inf. Proces. Syst., vol. 29, 2016, pp. 1–9. Data Mining, Aug. 2016, pp. 855–864.
[77] J. Bruna, W. Zaremba, A. Szlam, and Y. LeCun, ‘‘Spectral networks [100] J. Ye, J. Zhao, K. Ye, and C. Xu, ‘‘How to build a graph-based deep
and deep locally connected networks on graphs,’’ in Proc. 2nd Int. Conf. learning architecture in traffic domain: A survey,’’ IEEE Trans. Intell.
Learn. Represent., Apr. 2014, pp. 1–14. Transp. Syst., vol. 23, no. 5, pp. 3904–3924, May 2022.
[101] W. Jiang and J. Luo, ‘‘Graph neural network for traffic forecasting: YOONJIN YOON (Member, IEEE) received the
A survey,’’ Expert Syst. Appl., vol. 207, Nov. 2022, Art. no. 117921. B.S. degree in mathematics from Seoul National
[102] A. van den Oord, N. Kalchbrenner, L. Espeholt, k. kavukcuoglu, University, Seoul, South Korea, in 1996, the dual
O. Vinyals, and A. Graves, ‘‘Conditional image generation with M.S. degree in computer science and in man-
PixelCNN decoders,’’ in Proc. Adv. Neural Inf. Process. Syst., vol. 29, agement science and engineering from Stanford
Dec. 2016, pp. 1–9. University, Stanford, CA, USA, in 2000 and 2002,
[103] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega, and P. Vandergheynst, respectively, and the Ph.D. degree in civil and
‘‘The emerging field of signal processing on graphs: Extending high- environmental engineering from the University
dimensional data analysis to networks and other irregular domains,’’ IEEE
of California at Berkeley, Berkeley, CA, USA,
Signal Process. Mag., vol. 30, no. 3, pp. 83–98, May 2013.
in 2010. Since 2011, she has been an Assistant
[104] S. M. Pincus, ‘‘Approximate entropy as a measure of system complexity,’’
Proc. Nat. Acad. Sci. USA, vol. 88, no. 6, pp. 2297–2301, Mar. 1991. Professor with the Department of Civil and Environmental Engineering,
[105] M. T. Ribeiro, S. Singh, and C. Guestrin, ‘‘‘Why should I trust Korea Advanced Institute of Science and Technology (KAIST), Daejeon,
you?’ Explaining the predictions of any classifier,’’ in Proc. 22nd South Korea. Before arriving at KAIST, she was a Graduate Student
ACM SIGKDD Int. Conf. Knowl. Discov. Data Mining, Aug. 2016, Researcher with the National Center of Excellence in Air Transportation
pp. 1135–1144. Operations Research (NeXTOR), University of California at Berkeley,
[106] S. M. Lundberg and S.-I. Lee, ‘‘A unified approach to interpreting model from 2005 to 2010. Her previous research experience includes as a Research
predictions,’’ in Proc. Adv. Neural Inf. Process. Sys., vol. 30, Dec. 2017, Assistant with the Artificial Intelligence Center, SRI International, Menlo
pp. 1–10. Park, CA, USA, in 1999, and the Center of Reliability Computing, Stanford
[107] A. Barredo-Arrieta, I. Laña, and J. Del Ser, ‘‘What lies beneath: A note on University, Stanford, CA, USA, from 2000 to 2002. Her research interests
the explainability of black-box machine learning models for road traffic include the traffic management of both manned and unmanned vehicles
forecasting,’’ in Proc. IEEE Intell. Transp. Syst. Conf. (ITSC), Oct. 2019, using stochastic optimization. Her recent research efforts include data-driven
pp. 2232–2237. driving behavior analysis, and autonomous vehicle traffic flow management
using large-scale driving data.