Computer Science ›› 2023, Vol. 50 ›› Issue (9): 318-330.doi: 10.11896/jsjkx.221000064

• Computer Network • Previous Articles     Next Articles

EGCN-CeDML:A Distributed Machine Learning Framework for Vehicle Driving Behavior Prediction

LI Ke1,2,3, YANG Ling1, ZHAO Yanbo4, CHEN Yonglong4, LUO Shouxi1,3   

  1. 1 School of Computer and Artificial Intelligence,Southwest Jiaotong University,Chengdu 611756,China
    2 Engineering Research Center of Sustainable Urban Intelligent Transportation,Ministry of Education,Chengdu 611756,China
    3 Sichuan Provincial Key Laboratory of Network and Communication Technology,Chengdu 611756,China
    4 Leeds Joint School,Southwest Jiaotong University,Chengdu 611756,China
  • Received:2022-10-10 Revised:2023-05-04 Online:2023-09-15 Published:2023-09-01
  • About author:LI Ke,born in 1985,Ph.D,lecturer,is a member of China Computer Federation.Her main research interests include intelligent transportation system and Internet of vehicles.
  • Supported by:
    National Natural Science Foundation of China(62202392,61731017),Project of Network and Data Security Key Laboratory in Sichuan Province(NDS2022-1) and Natural Science Foundation of Sichuan Province,China(2023NSFSC0459,2022NSFSC0944).

Abstract: In large-scale dynamic traffic scenarios,predicting vehicle driving behavior quickly and accurately is one of the most challenging issues in the field of intelligent traffic driving.The prediction of vehicle driving behavior should consider not only the efficiency of communication,but also the historical vehicle trajectory and the interaction between vehicles.Considering the above factors,this paper proposes a communication-efficient distributed machine learning framework based on edge-enhanced graph convolutional neural networks(EGCN-CeDML).Compared with the centralized prediction framework on a single device,EGCN-CeDML is a communication-efficient distributed machine learning framework,which does not need to transmit all the raw data to the cloud server,and directly stores,processes,and computes user data locally.This way of training neural networks on multiple edge devices relieves the pressure of centralized training neural networks,reduces the amount of transmitted data and communication latency,improves data processing efficiency,and preserves user privacy to a certain extent.EGCN-LSTM deployed on each edge device utilizes the edge-enhanced attention mechanism and the feature transfer mechanism of the graph convolutional neural network to promptly extract and transfer the interaction information between vehicles when the number of surrounding vehicles increases to more than a dozen,ensuring more accurate prediction performance and lower time complexity.In addition to vehicle driving behavior prediction,each edge device can flexibly control the type and scale of the neural network according to its own computing and storage capabilities,under the premise of ensuring the performance of the neural network,which is suitable for different application scenarios.The experimental results of EGCN-CeDML on public dataset NGSIMshow that the amount of data to be transmitted by only accounts for 21.56% of the centralized training.And the calculation time and prediction performance of EGCN-CeDML are better than those of previous models regardless of traffic complexity,with an accuracy rate of 0.939 1,a recall rate of 0.955 7,and an F1 score of 0.947 3.When the prediction time is one second,the prediction accuracy reaches 91.21%.Even if the number of vehicles increases,the algorithm maintains a low time complexity and is stable within 0.1 seconds.

Key words: Vehicle driving behavior prediction, Graph convolutional network, Edge enhancement, Attention mechanism, Distributed machine learning

CLC Number: 

  • TP393
[1]XING E P,HO Q,DAI W,et al.Petuum:A new platform for distributed machine learning on big data[C]//Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.2015:1335-1344.
[2]ANDREW G.Bring HPC techniques to deep learning[EB/OL].https://xueshu.baidu.com/bringing-hpc-techniques-deep-learning.
[3]MA Y,ZHU X,ZHANG S,et al.Trafficpredict:Trajectory prediction for heterogeneous traffic-agents[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2019:6120-6127.
[4]AMDAHL G M.Validity of the single processor approach toachieving large scale computing capabilities[C]//Spring Joint Computer Conference.1967:483-485.
[5]GONG L,CHENG Q.Exploiting edge features for graph neural networks[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:9211-9219.
[6]LEFÈVRE S,VASQUEZ D,LAUGIER C.A survey on motion prediction and risk assessment for intelligent vehicles[J].ROBOMECH Journal,2014,1(1):1-14.
[7]LI Z,GONG J,LU C,et al.Interactive Behavior Prediction for Heterogeneous Traffic Participants in the Urban Road:A Graph-Neural-Network-Based Multitask Learning Framework[J].IEEE/ASME Transactions on Mechatronics,2021,26(3):1339-1349.
[8]KUMAR P,PERROLLAZ M,LEFEVRE S,et al.Learning-based approach for online lane change intention prediction[C]//2013 IEEE Intelligent Vehicles Symposium(IV).IEEE,2013:797-802.
[9]HE G,LI X,LV Y,et al.Probabilistic intention prediction and trajectory generation based on dynamic bayesian networks[C]//2019 Chinese Automation Congress(CAC).IEEE,2019:2646-2651.
[10]AOUDE G S,DESARAJU V R,STEPHENS L H,et al.Driver behavior classification at intersections and validation on large naturalistic data set[J].IEEE Transactions on Intelligent Transportation Systems,2012,13(2):724-736.
[11]DEO N,TRIVEDI M M.Convolutional social pooling for vehicle trajectory prediction[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops.2018:1468-1476.
[12]GAO J,SUN C,ZHAO H,et al.Vectornet:Encoding hd maps and agent dynamics from vectorized representation[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2020:11525-11533.
[13]BUCILA C,CARUANA R,NICULESCU-MIZIL A.ModelCompression[C]//ACM SIGKDD International Conference on Knowledge Discovery and Data Mining(KDD'06).2006.
[14]HINTON G,VINYALS O,DEAN J.Distilling the knowledge in a neural network[J].arXiv:1503.02531,2015.
[15]SCARSELLI F,GORI M,TSOI A C,et al.The graph neural network model[J].IEEE Transactions on Neural Networks,2008,20(1):61-80.
[16]WANG X,JI H,SHI C,et al.Heterogeneous graph attention network[C]//The World Wide Web Conference.2019:2022-2032.
[17]KRAJEWSKI R,BOCK J,KLOEKER L,et al.The highd dataset:A drone dataset of naturalistic vehicle trajectories on german highways for validation of highly automated driving systems[C]//2018 21st International Conference on Intelligent Transportation Systems(ITSC).IEEE,2018:2118-2125.
[18]GUO J.Research on Precise Hybrid Recommendation MethodBased on Feature Engineering and Efficient Gradient Boosting Tree Algorithm[D].Wuhan:Wuhan University of Science and Technology,2020.
[19]KRIZHEVSKY A,SUTSKEVER I,HINTON G E.Imagenetclassification with deep convolutional neural networks[J].Communications of the ACM,2017,60(6):84-90.
[20]RUMELHART D E,HINTON G E,WILLIAMS R J.Learning Representations by Back Propagating Errors[J].Nature,1986,323(6088):533-536.
[21]TEERAPITTAYANON S,MCDANEL B,KUNG H T.Distri-buted Deep Neural Networks Over the Cloud,the Edge and End Devices[C]//2017 IEEE 37th International Conference on Distributed Computing Systems(ICDCS).IEEE,2017:328-339.
[22]LECUN Y,BOTTOU L.Gradient-based learning applied to do-cument recognition[C]//Proceedings of the IEEE.1998:2278-2324.
[1] YI Liu, GENG Xinyu, BAI Jing. Hierarchical Multi-label Text Classification Algorithm Based on Parallel Convolutional Network Information Fusion [J]. Computer Science, 2023, 50(9): 278-286.
[2] LUO Yuanyuan, YANG Chunming, LI Bo, ZHANG Hui, ZHAO Xujian. Chinese Medical Named Entity Recognition Method Incorporating Machine ReadingComprehension [J]. Computer Science, 2023, 50(9): 287-294.
[3] WANG Jiahao, ZHONG Xin, LI Wenxiong, ZHAO Dexin. Human Activity Recognition with Meta-learning and Attention [J]. Computer Science, 2023, 50(8): 193-201.
[4] WANG Yu, WANG Zuchao, PAN Rui. Survey of DGA Domain Name Detection Based on Character Feature [J]. Computer Science, 2023, 50(8): 251-259.
[5] ZHANG Yian, YANG Ying, REN Gang, WANG Gang. Study on Multimodal Online Reviews Helpfulness Prediction Based on Attention Mechanism [J]. Computer Science, 2023, 50(8): 37-44.
[6] TENG Sihang, WANG Lie, LI Ya. Non-autoregressive Transformer Chinese Speech Recognition Incorporating Pronunciation- Character Representation Conversion [J]. Computer Science, 2023, 50(8): 111-117.
[7] YAN Mingqiang, YU Pengfei, LI Haiyan, LI Hongsong. Arbitrary Image Style Transfer with Consistent Semantic Style [J]. Computer Science, 2023, 50(7): 129-136.
[8] GAO Xiang, TANG Jiqiang, ZHU Junwu, LIANG Mingxuan, LI Yang. Study on Named Entity Recognition Method Based on Knowledge Graph Enhancement [J]. Computer Science, 2023, 50(6A): 220700153-6.
[9] ZHANG Tao, CHENG Yifei, SUN Xinxu. Graph Attention Networks Based on Causal Inference [J]. Computer Science, 2023, 50(6A): 220600230-9.
[10] CUI Lin, CUI Chenlu, LIU Zhengwei, XUE Kai. Speech Emotion Recognition Based on Improved MFCC and Parallel Hybrid Model [J]. Computer Science, 2023, 50(6A): 220800211-7.
[11] DUAN Jianyong, YANG Xiao, WANG Hao, HE Li, LI Xin. Document-level Relation Extraction of Graph Attention Convolutional Network Based onInter-sentence Information [J]. Computer Science, 2023, 50(6A): 220800189-6.
[12] YANG Xing, SONG Lingling, WANG Shihui. Remote Sensing Image Classification Based on Improved ResNeXt Network Structure [J]. Computer Science, 2023, 50(6A): 220100158-6.
[13] ZHANG Shunyao, LI Huawang, ZHANG Yonghe, WANG Xinyu, DING Guopeng. Image Retrieval Based on Independent Attention Mechanism [J]. Computer Science, 2023, 50(6A): 220300092-6.
[14] LIU Haowei, YAO Jingchi, LIU Bo, BI Xiuli, XIAO Bin. Two-stage Method for Restoration of Heritage Images Based on Muti-scale Attention Mechanism [J]. Computer Science, 2023, 50(6A): 220600129-8.
[15] LI Fan, JIA Dongli, YAO Yumin, TU Jun. Graph Neural Network Few Shot Image Classification Network Based on Residual and Self-attention Mechanism [J]. Computer Science, 2023, 50(6A): 220500104-5.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!