Channel attention-based spatial-temporal graph neural networks for traffic prediction
Data Technologies and Applications
ISSN: 2514-9288
Article publication date: 3 May 2023
Issue publication date: 29 January 2024
Abstract
Purpose
Traffic flow prediction has always been a top priority of intelligent transportation systems. There are many mature methods for short-term traffic flow prediction. However, the existing methods are often insufficient in capturing long-term spatial-temporal dependencies. To predict long-term dependencies more accurately, in this paper, a new and more effective traffic flow prediction model is proposed.
Design/methodology/approach
This paper proposes a new and more effective traffic flow prediction model, named channel attention-based spatial-temporal graph neural networks. A graph convolutional network is used to extract local spatial-temporal correlations, a channel attention mechanism is used to enhance the influence of nearby spatial-temporal dependencies on decision-making and a transformer mechanism is used to capture long-term dependencies.
Findings
The proposed model is applied to two common highway datasets: METR-LA collected in Los Angeles and PEMS-BAY collected in the California Bay Area. This model outperforms the other five in terms of performance on three performance metrics a popular model.
Originality/value
(1) Based on the spatial-temporal synchronization graph convolution module, a spatial-temporal channel attention module is designed to increase the influence of proximity dependence on decision-making by enhancing or suppressing different channels. (2) To better capture long-term dependencies, the transformer module is introduced.
Keywords
Citation
Wang, B., Gao, F., Tong, L., Zhang, Q. and Zhu, S. (2024), "Channel attention-based spatial-temporal graph neural networks for traffic prediction", Data Technologies and Applications, Vol. 58 No. 1, pp. 81-94. https://doi.org/10.1108/DTA-09-2022-0378
Publisher
:Emerald Publishing Limited
Copyright © 2023, Emerald Publishing Limited