Computer Science ›› 2023, Vol. 50 ›› Issue (4): 40-46.doi: 10.11896/jsjkx.220200079

• Database & Big Data & Data Science • Previous Articles     Next Articles

Short-time Traffic Flow Forecasting Based on Multi-adjacent Graph and Multi-head Attention Mechanism

YIN Heng1, ZHANG Fan1,2, LI Tianrui1,2,3   

  1. 1 School of Computing and Artificial Intelligence,Southwest Jiaotong University,Chengdu 611756,China
    2 Manufacturing Industry Chains Collaboration and Information Support Technology Key Laboratory of Sichuan Province,Chengdu 611756,China
    3 National Engineering Laboratory of Integrated Transportation Big Data Application Technology,Chengdu 611756,China
  • Received:2022-02-16 Revised:2022-09-05 Online:2023-04-15 Published:2023-04-06
  • About author:YIN Heng,born in 1997,postgraduate.His main research interests include traffic flow prediction and urban computing.
    ZHANG Fan,born in 1989,Ph.D,research assistant.His main research interests include fault diagnosis,pattern recognition and data fusion.
  • Supported by:
    National Natural Science Foundation of China(61773324) and Sichuan Key R&D Project,China(2020YFG0035).

Abstract: Traffic flow forecasting is the cornerstone of many applications in transportation which has a great importance in smart city system.The difficulty of this task is how to effectively model the temporal and spatial dependence.Existing methods usually use GNN to model temporal correlation and CNN or RNN to model temporal correlation.When modeling the spatial correlation,only the adjacency matrix is applied to model local relationships while ignoring global spatial information.However,there are some roads in the entire road network whose surrounding structures are similar,and these roads carry similar functions in the road network.Therefore,the characteristics of these similar roads can also be used as the basis for traffic prediction.This paper proposes a traffic flow forecasting model based on multi-adjacent matrix and multi-head attention mechanism.It includes:1)the node2vec algorithm is applied to calculate the vector representation of the road in road network,and the similarity matrix is calculated through the threshold for graph convolution operation to extract global spatial information;2)the multi-channel self-attention mechanism is used to mine the spatial and temporal features of the model.Experiments on public datasets PEMS04 and PEMS08 demonstrate the proposed model’s effectiveness.Its accuracy is improved compared with the mainstream models.

Key words: Traffic forecasting, Dilated convolution, Spatial-Temporal network, Attention mechanism, Node embedding

CLC Number: 

  • TP391
[1]AGARAP A F M.A neural network architecture combining gated recurrent unit(GRU)and support vector machine(SVM)for intrusion detection in network traffic data[C]//Proceedings of the 2018 10th International Conference on Machine Learning and Computing.2018:26-30.
[2]WILLIAMS B M,HOEL L A.Modeling and forecasting vehicular traffic flow as a seasonal arima process:Theoretical basis and empirical results[J].Journal of Transportation Engineering,2003,129(6):664-672.
[3]LESHEM G,RITOV Y.Traffic flow prediction using adaboost algorithm with random forests as a weak learner[C]//Procee-dings of World Academy of Science,Engineering and Technology.2007:193-198.
[4]LECUN Y,BENGIO Y,HINTON G.Deep learning[J].Nature,2015,521(7553):436-444.
[5]CUI Z,KE R,PU Z,et al.Deep bidirectional and unidirectional LSTM recurrent neural network for network-wide traffic speed prediction[J].arXiv:1801.02143,2018.
[6]ZHANG X H,FENG A M.Short-Time traffic flow prediction based on and LSTM network[J].Application Research of Computers,2021,41(1):225-230.
[7]SHI X J,CHEN Z,WANG H,et al.Convolutional LSTM network:A machine learning approach for precipitation nowcasting[C]//Advances inNeural Information Processing Systems.2015:802-810.
[8]ZHANG J,ZHENG Y,QI D.Deep spatio-temporal residual networks for citywide crowd flows prediction[C]//Proceedings of Thirty-first AAAI Conference on Artificial Intelligence.2017:1566-1661.
[9]ESTRACH J B,ZAREMBA W,SZLAM A,et al.Spectral networks and deep locally connected networks on graphs[C]//Proceedings of 2nd International Conference on Learning Representations.2014:1-14.
[10]SEO Y,DEFFERRARD M,VANDERGHEYNST P,et al.Struc-tured sequence modeling with graph convolutional recurrent networks[C]//Proceedings of International Conference on Neural Information Processing.Cham:Springer,2018:362-373.
[11]LI Y,YU R,SHAHABI C,et al.Diffusion convolutional recurrent neural network:Data-driven traffic forecasting[C]//Proceedings of International Conference on Learning Representation.2018:1-16.
[12]SUTSKEVER I,VINYALS O,LE Q V.Sequence to sequencelearning with neural networks[C]//Advances inNeural Information Processing Systems.2014:3104-3112.
[13]XIONG T,QI Y,ZHANG W B.Short-Time traffic flow prediction in road network based on DCGRU-RF model[J].Computer Science,2020,47(5):84-89.
[14]YU B,YIN H,ZHU Z.Spatio-temporal graph convolutional networks:A deep learning framework for traffic forecasting[J].arXiv:1709.04875,2017.
[15]GUO S,LIN Y,FENG N,et al.Attention based spatial-temporal graph convolutional networks for traffic flow forecasting[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2019:922-929.
[16]WU Z,PAN S,LONG G,et al.Graph wavenet for deep spatial-temporal graph modeling[J].arXiv:1906.00121,2019.
[17]YU F,KOLTUN V.Multi-scale context aggregation by dilated convolutions[J].arXiv:1511.07122,2015.
[18]LI H,WANG F,XIE S Y,et al.A traffic prediction model with dual autoregressive components based on improved graph wave network[J].Computer Science,2021,48(S2):159-165,183.
[19]SONG C,LIN Y,GUO S,et al.Spatial-temporal synchronousgraph convolutional networks:A new framework for spatial-temporal network data forecasting[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2020:914-921.
[20]PEROZZI B,AL-RFOU R,SKIENA S.Deepwalk:Online lear-ning of social representations[C]//Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Disco-very and Data Mining.2014:701-710.
[21]MIKOLOV T,SUTSKEVER I,CHEN K,et al.Distributed representations of words and phrases and their compositionality[C]//Advances in Neural Information Processing Systems.2013:3111-3119.
[22]TANG J,QU M,WANG M,et al.Line:Large-scale information network embedding[C]//Proceedings of the 24th International Conference on World Wide Web.2015:1067-1077.
[23]GROVER A,LESKOVEC J.node2vec:Scalable feature learning for networks[C]//Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mi-ning.2016:855-864.
[24]VASWANI A,SHAZEER N,PARMAR N,et al.Attention isall you need[C]//Advances inNeural Information Processing Systems.2017:5998-6008.
[25]WEISS M,BONNEL P,HUMMEL R,et al.Analyzing on-road emissions of light-duty vehicles with Portable Emission Mea-surement Systems(PEMS)[R].JRC Scientific and Technical Reports,EUR,2011.
[26]DEFFERRARD M,BRESSON X,VANDERGHEYNST P.Con-volutional neural networks on graphs with fast localized spectral filtering[J].Advances in Neural Information Processing Systems,2016,29:3844-3852.
[27]SRIVASTAVA N,HINTON G,KRIZHEVSKY A,et al.Dropout:a simple way to prevent neural networks from overfitting[J].Journal of Machine Learning Research,2014,15(1):1929-1958.
[1] YU Xingzhan, LU Tianliang, DU Yanhui, WANG Xirui, YANG Cheng. Android Malware Family Classification Method Based on Synthetic Image and Xception Improved Model [J]. Computer Science, 2023, 50(4): 351-358.
[2] HAN Xueming, JIA Caiyan, LI Xuanya, ZHANG Pengfei. Dual-attention Network Model on Propagation Tree Structures for Rumor Detection [J]. Computer Science, 2023, 50(4): 22-31.
[3] LUO Xiaohui, WU Yun, WANG Chenxing, YU Wenting. Sequential Recommendation Model Based on User’s Long and Short Term Preference [J]. Computer Science, 2023, 50(4): 47-55.
[4] WANG Yali, ZHANG Fan, YU Zeng, LI Tianrui. Aspect-level Sentiment Classification Based on Interactive Attention and Graph Convolutional Network [J]. Computer Science, 2023, 50(4): 196-203.
[5] LI Shuai, XU Bin, HAN Yike, LIAO Tongxin. SS-GCN:Aspect-based Sentiment Analysis Model with Affective Enhancement and Syntactic Enhancement [J]. Computer Science, 2023, 50(3): 3-11.
[6] CHEN Fuqiang, KOU Jiamin, SU Limin, LI Ke. Multi-information Optimized Entity Alignment Model Based on Graph Neural Network [J]. Computer Science, 2023, 50(3): 34-41.
[7] ZHOU Mingqiang, DAI Kailang, WU Quanwang, ZHU Qingsheng. Attention-aware Multi-channel Graph Convolutional Rating Prediction Model for Heterogeneous Information Networks [J]. Computer Science, 2023, 50(3): 129-138.
[8] QU Zhong, WANG Caiyun. Crack Detection of Concrete Pavement Based on Attention Mechanism and Lightweight DilatedConvolution [J]. Computer Science, 2023, 50(2): 231-236.
[9] LIU Luping, ZHOU Xin, CHEN Junjun, He Xiaohai, QING Linbo, WANG Meiling. Event Extraction Method Based on Conversational Machine Reading Comprehension Model [J]. Computer Science, 2023, 50(2): 275-284.
[10] ZOU Yunzhu, DU Shengdong, TENG Fei, LI Tianrui. Visual Question Answering Model Based on Multi-modal Deep Feature Fusion [J]. Computer Science, 2023, 50(2): 123-129.
[11] CAI Xiao, CEHN Zhihua, SHENG Bin. SPT:Swin Pyramid Transformer for Object Detection of Remote Sensing [J]. Computer Science, 2023, 50(1): 105-113.
[12] ZHANG Jingyuan, WANG Hongxia, HE Peisong. Multitask Transformer-based Network for Image Splicing Manipulation Detection [J]. Computer Science, 2023, 50(1): 114-122.
[13] LI Xuehui, ZHANG Yongjun, SHI Dianxi, XU Huachi, SHI Yanyan. AFTM:Anchor-free Object Tracking Method with Attention Features [J]. Computer Science, 2023, 50(1): 138-146.
[14] ZHAO Qian, ZHOU Dongming, YANG Hao, WANG Changchen. Image Deblurring Based on Residual Attention and Multi-feature Fusion [J]. Computer Science, 2023, 50(1): 147-155.
[15] ZHENG Cheng, MEI Liang, ZHAO Yiyan, ZHANG Suhang. Text Classification Method Based on Bidirectional Attention and Gated Graph Convolutional Networks [J]. Computer Science, 2023, 50(1): 221-228.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!