计算机科学 ›› 2024, Vol. 51 ›› Issue (6): 215-222.doi: 10.11896/jsjkx.230500085
李越豪1,2, 王邓江3, 鉴海防1, 王洪昌1,2, 程清华1,2
LI Yuehao1,2, WANG Dengjiang3, JIAN Haifang1, WANG Hongchang1,2, CHENG Qinghua1,2
摘要: 激光雷达工作环境中的光束衰减和目标遮挡会导致输出点云出现远端稀疏的问题,从而引起基于激光雷达的3D目标检测算法的检测精度随距离衰减的现象。针对这一问题,提出了一种基于鸟瞰图视角(BEV)空间内目标占位预测的激光-毫米波雷达融合目标检测算法。首先提出了一种简化的BEV占位预测子网络,用于生成位置相关的毫米波雷达特征,同时有助于解决毫米波雷达数据稀疏带来的网络收敛困难的问题。然后,为了实现跨模态特征融合,设计了一种基于BEV空间特征关联的多尺度激光-毫米波雷达特征融合层结构。在nuScenes数据集上进行实验,结果表明,所提出的毫米波雷达分支网络的平均检测精度(mAP)达到21.6%,推理时间为8.3ms。在加入融合层结构后,多模态检测算法较基线算法CenterPoint的mAP提升了2.9%,同时增加的额外推理时间开销仅为8.6ms,在距离传感器30m位置处,多模态算法对于nuScenes数据集中10个类别的检测精度达成率分别较CenterPoint提升了2.1%~16.0%。
中图分类号:
[1]KIM S H,HWANG Y.A survey on deep learning based me-thods and datasets for monocular 3d object detection[J].Electro-nics,2021,10(4):517. [2]HAN B,ZHANG X Y,REN S.Survey of convolution operations based on 3D point clouds[J].Journal of Computer Research and Development,2023,60(4):873-902. [3]CAESAR H,BANKITI V,LANG A H,et al.Nuscenes:A multimodal dataset for autonomous driving[C]//Proceedings of the 33rd IEEE Conference on Computer Vision and Pattern Recognition.Piscataway,NJ:IEEE,2020:11618-11628. [4]QI C R,SU H,MO K,et al.PointNet:Deep learning on point sets for 3D classification and segmentation[C]//Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Re-cognition.Piscataway,NJ:IEEE,2017:55-63. [5]QI C R,YI L,SU H,et al.PointNet++:Deep hierarchical feature learning on point sets in a metric space[C]//Proceedings of the 31st Int Conference on Neural Information Processing Systems.Cambridge,MA:MIT Press,2017:5105-5114. [6]SHI S,WANG X,LI H.PointRCNN:3D object proposal generation and detection from point cloud[C]//Proceedings of the 32nd IEEE Conference on Computer Vision and Pattern Recognition.Piscataway,NJ:IEEE,2019:770-779. [7]ZHANG Y,HU Q,XU G,et al.Not all points are equal:Lear-ning highly efficient point-based detectors for 3D LiDAR point clouds[C]//Proceedings of the 35th IEEE Conference on Computer Vision and PatternRecognition.Piscata-way,NJ:IEEE,2022:18931-18940. [8]YA Z Y,TUZEL O.VoxelNet:End-to-end learning for point cloud based 3D object detection[C]//Proceedings of the 31st IEEE Conference on Computer Vision andPattern Recognition.Piscataway,NJ:IEEE,2018:4490-4499. [9]YAN Y,MAO Y,LI B.Second:Sparsely embedded convolu-tional detection[J].Sensors,2018,18(10):3337. [10]YIN T,ZHOU X,KRÄHENBÜHL P.Center-based 3D object detection and tracking[C]//Proceedings of the 35th IEEE Conference on Computer Vision and Pattern Recognition.Pisca-taway,NJ:IEEE,2021:11779-11788. [11]DENG J,SHI S,LI P,et al.Voxel R-CNN:Towards high performance voxel-based 3D object detection.[C]//Proceedings of the 35th AAAI Conference on Artificial Intelligence.Menlo Park,CA:AAAI,2021:1201-1209. [12]LANG A H,VORA S,CAESAR H,et al.Pointpillars:Fast encoders for object detection from point clouds[C]//Proceedings of the 32nd IEEE Conference on Computer Vision and Pattern Recognition.Piscataway,NJ:IEEE,2019:12689-12697. [13]LI Y,SHEN T,ZENG K.3D object detection based on attention fusion of millimeter wave radar point cloud and visual information disparity features[J].Journal of Optoelectronics·Laser,2023,34(1):26-33. [14]DREHER M,ERÇELIK E,BÄNZIGER T,et al.Radar-based 2D car detection using deep neural networks[C]//Proceedings of the 23rd IEEE International Conference on Intelligent Transportation Systems Piscataway.NJ:IEEE,2020:1-8. [15]YOLOv3:An Incremental Improvement[J].arXiv:1804.02767,2018. [16]QI C R,LIU W,WU C,et al.Frustum pointnets for 3D object detection from RGB-D data[C]//Proceedings of the 31st IEEE Conference on Computer Vision and Pattern Recognition.Piscataway,NJ:IEEE,2018:918-927. [17]MICHAEL U,SASCHA B,DANIEL K,et al.Improved orientation estimation and detection with hybrid object detection networks for automotive radar[C]//Proceedings of the 23rd IEEE International Conference on Intelligent Transportation Systems Piscataway.NJ:IEEE,2022:918-927. [18]THOMAS H,QI C R,DESCHAUD J E,et al.KPConv:Flexible and deformable convolution for point clouds[C]//Proceedings of the 17th IEEE International Conference on Computer Vision.Piscataway,NJ:IEEE,2019:6410-6419. [19]NABATI R,QI H.RRPN:Radar region proposal network for object detection in autonomous vehicles[C]//Proceedings of the IEEE International Conference on Image Processing.Pisca-taway,NJ:IEEE,2019:3093-3097. [20]CHADWICK S,MADDETN W,NEWMAN P.Distant vehicle detection using radar and vision[C]//Proceedings of the 2019 International Conference on Robotics and Automation.Pisca-taway.NJ:IEEE,2019:8311-8317. [21]NOBIS F,GEISSLINGER M,WEBER M,et al.A deep learning-based radar and camera sensor fusion architecture for object detection[C]//Proceedings of the 2019 IEEE Symposium on Sensor Data Fusion:Trends,Solutions,Applications.Piscataway,NJ:IEEE,2019:1-7. [22]BIJELIC M,GRUBER T,MANNAN F,et al.Seeing through fog without seeing fog:Deep multimodal sensor fusion in unseen adverse weather[C]//Proceedings of the 34th IEEE Conference on Computer Vision and Pattern Recognition.Piscataway,NJ:IEEE,2020:11679-11689. [23]TAO H Z,YI N S,JUN J C,et al.Behind the curtain:Learning occluded shapes for 3D object detection[C]//Proceedings of the 36th AAAI Conference on Artificial Intelligence.Menlo Park,CA:AAAI,2022:11910-11919. |
|