Computer Science ›› 2023, Vol. 50 ›› Issue (1): 131-137.doi: 10.11896/jsjkx.211100097
• Computer Graphics & Multimedia • Previous Articles Next Articles
CHEN Yunfang, LU Yangyang, ZHOU Xin, ZHANG Wei
CLC Number:
[1]LEE B,ERDENEE E,JIN S,et al.Multi-class multi-objecttracking using changing point detection[C]//European Confe-rence on Computer Vision.Cham:Springer,2016:68-83. [2]WOJKE N,BEWLEY A,PAULUS D.Simple online and real-time tracking with a deep association metric[C]//2017 IEEE International Conference on Image Processing(ICIP).IEEE,2017:3645-3649. [3]FANG K,XIANG Y,LI X,et al.Recurrent autoregressive networks for online multi-object tracking[C]//2018 IEEE Winter Conference on Applications of Computer Vision(WACV).IEEE,2018:466-475. [4]FARHADI A,REDMON J.Yolov3:An incremental improve-ment[C]//Computer Vision and Pattern Recognition.Berlin/Heidelberg,Germany:Springer,2018:1804-02. [5]REN S,HE K,GIRSHICK R,et al.Faster r-cnn:Towards real-time object detection with region proposal networks[J].Advances in Neural Information Processing Systems,2015,28:91-99. [6]GONG X,LE Z C,WNAG H,et al.Survey of Data Association Technology in Multi-target Tracking[J].Computer Science,2020,47(10):136-144. [7]SADEGHIAN A,ALAHI A,SAVARESE S.Tracking the untrackable:Learning to track multiple cues with long-term dependencies[C]//Proceedings of the IEEE International Confe-rence on Computer Vision.2017:300-311. [8]REZATOFIGHI S H,MILAN A,ZHANG Z,et al.Joint probabilistic data association revisited[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:3047-3055. [9]KIM C,LI F,CIPTADI A,et al.Multiple hypothesis trackingrevisited[C]//Proceedings of the IEEE International Confe-rence on Computer Vision.2015:4696-4704. [10]LEAL-TAIXÉ L,CANTON-FERRER C,SCHINDLER K.Learning by tracking:Siamese CNN for robust target association[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops.2016:33-40. [11]SUN S J,AKHTAR N,SONG H S,et al.Deep affinity network for multiple object tracking[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2019,43(1):104-119. [12]MAHMOUDI N,AHADI S M,RAHMATI M.Multi-targettracking using CNN-based features:CNNMTT[J].Multimedia Tools and Applications,2019,78(6):7077-7096. [13]BAE S H,YOON K J.Confidence-based data association and discriminative deep appearance learning for robust online multi-object tracking[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2017,40(3):595-610. [14]BERGMANN P,MEINHARDT T,LEAL-TAIXE L.Tracking without bells and whistles[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision.2019:941-951. [15]WANG Z,ZHENG L,LIU Y,et al.Towards real-time multi-object tracking[C]//Computer Vision-ECCV 2020:16th European Conference,Glasgow,UK(Part XI 16).Springer International Publishing,2020:107-122. [16]LIN T Y,DOLLÁR P,GIRSHICK R,et al.Feature pyramidnetworks for object detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:2117-2125. [17]ZHANG Y,WANG C,WANG X,et al.A simple baseline for multi-object tracking[J].arXiv:2004.01888,2020. [18]CHEN L,AI H,ZHUANG Z,et al.Real-time multiple people tracking with deeply learned candidate selection and person re-identification[C]//2018 IEEE International Conference on Multimedia and Expo(ICME).IEEE,2018:1-6. [19]KUHN H W.The Hungarian method for the assignment problem[J].Naval Research Logistics Quarterly,1955,2(1/2):83-97. [20]HE K,ZHANG X,REN S,et al.Deep residual learning forimage recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770-778. [21]LIU W,ANGUELOV D,ERHAN D,et al.Ssd:Single shot multibox detector[C]//European Conference on Computer vision.Cham:Springer,2016:21-37. [22]FELZENSZWALB P F,GIRSHICK R B,MCALLESTER D,et al.Object detection with discriminatively trained part-based models[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2009,32(9):1627-1645. [23]BERNARDIN K,STIEFELHAGEN R.Evaluating multiple object tracking performance:the clear mot metrics[J].EURASIP Journal on Image and Video Processing,2008,2008:1-10. [24]HE K,ZHANG X,REN S,et al.Delving deep into rectifiers:Surpassing human-level performance on imagenet classification[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:1026-1034. [25]YU F,LI W,LI Q,et al.Poi:Multiple object tracking with high performance detection and appearance feature[C]//European Conference on Computer Vision.Cham:Springer,2016:36-42. [26]PENG J,WANG C,WAN F,et al.Chained-tracker:Chainingpaired attentive regression results for end-to-end joint multiple-object detection and tracking[C]//European Conference on Computer Vision.Cham:Springer,2020:145-161. [27]KIM C,LI F,REHG J M.Multi-object tracking with neural gating using bilinear lstm[C]//Proceedings of the European Conference on Computer Vision(ECCV).2018:200-215. [28]CHEN J,SHENG H,ZHANG Y,et al.Enhancing detectionmodel for multiple hypothesis tracking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops.2017:18-27. [29]ZHU J,YANG H,LIU N,et al.Online multi-object tracking with dual matching attention networks[C]//Proceedings of the European Conference on Computer Vision(ECCV).2018:366-382. [30]CHOI W.Near-online multi-target tracking with aggregated local flow descriptor[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:3029-3037. [31]KEUPER M,TANG S,ANDRES B,et al.Motion segmentation &multiple object tracking by correlation co-clustering[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2018,42(1):140-153. |
[1] | MA Rui-xin, LI Ze-yang, CHEN Zhi-kui, ZHAO Liang. Review of Reasoning on Knowledge Graph [J]. Computer Science, 2022, 49(6A): 74-85. |
[2] | LI Sun, CAO Feng. Analysis and Trend Research of End-to-End Framework Model of Intelligent Speech Technology [J]. Computer Science, 2022, 49(6A): 331-336. |
[3] | YANG Run-yan, CHENG Gao-feng, LIU Jian. Study on Keyword Search Framework Based on End-to-End Automatic Speech Recognition [J]. Computer Science, 2022, 49(1): 53-58. |
[4] | ZHANG Peng, WANG Xin-qing, XIAO Yi, DUAN Bao-guo, XU Hong-hui. Real-time Binocular Depth Estimation Algorithm Based on Semantic Edge Drive [J]. Computer Science, 2021, 48(9): 216-222. |
[5] | LIU Dong, WANG Ye-fei, LIN Jian-ping, MA Hai-chuan, YANG Run-yu. Advances in End-to-End Optimized Image Compression Technologies [J]. Computer Science, 2021, 48(3): 1-8. |
[6] | JIANG Qi, SU Wei, XIE Ying, ZHOUHONG An-ping, ZHANG Jiu-wen, CAI Chuan. End-to-End Chinese-Braille Automatic Conversion Based on Transformer [J]. Computer Science, 2021, 48(11A): 136-141. |
[7] | LIU Yan, QIN Pin-le, ZENG Jian-chao. Multi-object Tracking Algorithm Based on YOLOv3 and Hierarchical Data Association [J]. Computer Science, 2021, 48(11A): 370-375. |
[8] | HUA Ming, LI Dong-dong, WANG Zhe, GAO Da-qi. End-to-End Speaker Recognition Based on Frame-level Features [J]. Computer Science, 2020, 47(10): 169-173. |
[9] | HUA Zhen, ZHANG Hai-cheng, LI Jin-jiang. End-to-end Image Super Resolution Based on Residuals [J]. Computer Science, 2019, 46(6): 246-255. |
[10] | WANG Zheng-ning, ZHOU Yang, LV Xia, ZENG Fan-wei, ZHANG Xiang, ZHANG Feng-jun. Improved MDP Tracking Method by Combining 2D and 3D Information [J]. Computer Science, 2019, 46(3): 97-102. |
[11] | ZHAO Guang-hui, ZHUO Song, XU Xiao-long. Multi-object Tracking Algorithm Based on Kalman Filter [J]. Computer Science, 2018, 45(8): 253-257. |
[12] | MA Ding, ZHUANG Lei and LAN Ju-long. Research on End-to-End Model of Reconfigurable Information Communication Basal Network [J]. Computer Science, 2017, 44(6): 114-120. |
[13] | GUO Zi-rong, ZENG Hua-xin and DOU Jun. Simple and Smoothed Fair Round Robin Scheduling Algorithm [J]. Computer Science, 2016, 43(1): 122-127. |
[14] | JIANG Nan and HE Yuan-zhi. ACO Based Traffic Classified Routing Algorithm in Distributed Satellite Cluster Network [J]. Computer Science, 2015, 42(10): 95-100. |
[15] | LIU Zhi-zhong,WANG Yong,HE Yi-hui and PENG Hui. QoS Aggregation for End-to-End QoS Requirement in Service Composition [J]. Computer Science, 2013, 40(Z6): 19-21. |
|