Computer Science ›› 2022, Vol. 49 ›› Issue (12): 236-243.doi: 10.11896/jsjkx.220600037

• Computer Graphics & Multimedia • Previous Articles     Next Articles

RGBT Object Tracking Based on High Rank Feature and Position Attention

YANG Lan-lan, WANG Wen-qi, WANG Fu-tian   

  1. School of Computer Science and Technology,Anhui University,Hefei 230000,China
  • Received:2022-06-06 Revised:2022-07-25 Published:2022-12-14
  • About author:YANG Lan-lan,born in 1994,postgra-duate.Her main research interests include RGBT object tracking and so on.WANG Fu-tian,born in 1981,Ph.D,professor.His main research interests include image processing,computer vision and edge computing.

Abstract: RGBT target tracking uses the advantages of two different modes of visible light(RGB) and thermal infrared(T) to solve the common modal limitation problem in single mode target tracking,so as to improve the performance of target tracking in complex environment.In the RGBT object tracking algorithm,the precise location of the object and the effective fusion of the two modalities are very important issues.In order to accurately locate the object and effectively fuse the two modalities,this paper proposes a new method to explore high-rank feature maps and introduce position attention for RGBT object tracking.The method first uses location attention to focus on the location information of the object according to the deep and shallow features of the backbone network,and then focuses on the importance of the features by exploring the high-rank feature maps before the fusion of the two modalities to guide the modal features fusion.In order to focus on the object location information,this paper uses the average pooling operation on the rows and columns.For the high-rank feature guidance module,this paper guides the fusion of feature maps according to the rank of the feature maps.In order to remove redundancy and noise and achieve more robust feature expression,the feature graph with small rank is deleted directly.Experimental results on two RGBT tracking benchmark data sets show that compared with other RGBT target tracking methods,the proposed method achieves better tracking results in accuracy and success rate.

Key words: RGBT object tracking, High rank feature, Object location information

CLC Number: 

  • TP18
[1]JI H X Y,LIANG P P,CHAI Y M,et al.Planar Object Tra-cking Algorithm Based on Key Points and Optical Flow[J].Computer Engineering,2021,47(4):234-240.
[2]DANELLJAN M,BHAT G,KHAN F S,et al.Atom:Accurate tracking by overlap maximization[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:4660-4669.
[3]WANG Q,ZHANG L,BERTINETTO L,et al.Fast online object tracking and segmentation:A unifying approach[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:1328-1338.
[4]WU Y,LIM J,YANG M H.Online object tracking:A benchmark[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2013:2411-2418.
[5]VALMADRE J,BERTINETTO L,HENRIQUES J F,et al.Long-term tracking in the wild:A benchmark[C]//Proceedings of the European Conference on Computer Vision(ECCV).2018:670-685.
[6]MARTIN D,GUSTAV H,FAHAD K,et al.Learning Spatially Regularized Correlation Filters for Visual Tracking[C]//ICCV.2015.
[7]MARTIN D,BHAT G,SHAHBAZ KHAN F,et al.Eco:Effi-cient convolution operators for tracking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:6638-6646.
[8]LI F,TIAN C,ZUO W,et al.Learning spatial-temporal regularized correlation filters for visual tracking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2018:4904-4913.
[9]LI G,YU Y.Deep contrast learning for salient object detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:478-487.
[10]WU A,ZHENG W S,YU H X,et al.RGB-infrared cross-moda-lity person re-identification[C]//Proceedings of the IEEE International Conference on Computer Vision.2017:5380-5389.
[11]BADRINARAYANAN V,KENDALL A,SEGNET R C.Adeep convolutional encoder-decoder architecture for image segmentation[J].arXiv:1511.00561,2015.
[12]LI C,XIA W,YAN Y,et al.Segmenting objects in day andnight:Edge-conditioned cnn for thermal image semantic segmentation[J].IEEE Transactions on Neural Networks and Learning Systems,2020,32(7):3069-3082.
[13]XU D,OUYANG W,RICCI E,et al.Learning cross-modal deep representations for robust pedestrian detection[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:5363-5371.
[14]GADE R,MOESLUND T B.Thermal cameras and applications:a survey[J].Machine Vision and Applications,2014,25(1):245-262.
[15]LI C,ZHAO N,LU Y,et al.Weighted sparse representationregularized graph learning for RGB-T object tracking[C]//Proceedings of the 25th ACM International Conference on Multimedia.2017:1856-1864.
[16]LAN X,YE M,ZHANG S,et al.Robust collaborative discriminative learning for RGB-infrared tracking[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2018:7008-7015.
[17]WANG Y,LI C,TANG J.Learning soft-consistent correlation filters for RGB-T object tracking[C]//Chinese Conference on Pattern Recognition and Computer Vision(PRCV).Cham:Springer,2018:295-306.
[18]LI C,ZHU C,HUANG Y,et al.Cross-modal ranking with soft consistency and noisy labels for robust RGB-T tracking[C]//Proceedings of the European Conference on Computer Vision(ECCV).2018:808-823.
[19]ZHANG L,DANELLJAN M,GONZALEZ-GARCIA A,et al.Multi-modal fusion for end-to-end rgb-t tracking[C]//Procee-dings of the IEEE/CVF International Conference on Computer Vision Workshops.2019:2252-2261.
[20]ZHANG H,ZHANG L,ZHUO L,et al.Object tracking inRGB-T videos using modal-aware attention network and competitive learning[J].Sensors,2020,20(2):393-411.
[21]WANG C,XU C,CUI Z,et al.Cross-modal pattern-propagation for RGB-T tracking[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2020:7064-7073.
[22]ZHANG P,ZHAO J,BO C,et al.Jointly modeling motion and appearance cues for robust RGB-T tracking[J].IEEE Transactions on Image Processing,2021,30:3335-3347.
[23]LI C,LIU L,LU A,et al.Challenge-aware RGBT tracking[C]//European Conference on Computer Vision.Cham:Sprin-ger,2020:222-237.
[24]LIN M,JI R,WANG Y,et al.Hrank:Filter pruning using high-rank feature map[C]//Proceedings of the IEEE/CVF Confe-rence on Computer Vision and Pattern Recognition.2020:1529-1538.
[25]HOU Q,ZHOU D,FENG J.Coordinate attention for efficient mobile network design[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2021:13713-13722.
[26]WOO S,PARK J,LEE J Y,et al.Cbam:Convolutional block attention module[C]//Proceedings of the European Conference on Computer Vision(ECCV).2018:3-19.
[27]HU J,SHEN L,ALBANIE S,et al.Squeeze-and-Excitation Networks[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2019,42(8):2011-2023.
[28]ZHU Y,LI C,TANG J,et al.Quality-aware feature aggregation network for robust RGBT tracking[J].IEEE Transactions on Intelligent Vehicles,2020,6(1):121-130.
[29]TANG Z,XU T,LI H,et al.Exploring Fusion Strategies for Accurate RGBT Visual Object Tracking[J].arXiv:2201.08673,2022.
[30]CHATFIELD K,SIMONYAN K,VEDALDI A,et al.Return of the devil in the details:Delving deep into convolutional nets[J].arXiv:1405.3531,2014.
[31]SIMONYAN K,ZISSERMAN A.Very deep convolutional networks for large-scale image recognition[J].arXiv:1409.1556,2014.
[32]BOTTOU L.Large-scale machine learning with stochastic gra-dient descent[C]//Proceedings of COMPSTAT.2010:177-186.
[33]NAM H,HAN B.Learning multi-domain convolutional neural networks for visual tracking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:4293-4302.
[34]LI C,CHENG H,HU S,et al.Learning collaborative sparse representation for grayscale-thermal tracking[J].IEEE Transa-ctions on Image Processing,2016,25(12):5743-5756.
[35]LI C,LIANG X,LU Y,et al.RGB-T object tracking:Benchmark and baseline[J].Pattern Recognition,2019,96:106977.
[36]PU S,SONG Y,MA C,et al.Deep attentive tracking via reciprocative learning[J].Advances in Neural Information Processing Systems,2018,31:1935-1945.
[37]GAO Y,LI C,ZHU Y,et al.Deep adaptive fusion network for high performance RGBT tracking[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops.2019:91-99.
[38]ZHU Y,LI C,LUO B,et al.Dense feature aggregation andpruning for rgbt tracking[C]//Proceedings of the 27th ACM International Conference on Multimedia.2019:465-472.
[39]JUNG I,SON J,BAEK M,et al.Real-time mdnet[C]//Procee-dings of the European Conference on Computer Vision(ECCV).2018:83-98.
[40]ZHANG Z,PENG H.Deeper and wider siamese networks for real-time visual tracking[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:4591-4600.
[41]DANELLJAN M,BHAT G,SHAHBAZ KHAN F,et al.Eco:Efficient convolution operators for tracking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:6638-6646.
[42]LI C,ZHAO N,LU Y,et al.Weighted sparse representation regularized graph learning for RGB-T object tracking[C]//Proceedings of the 25th ACM International Conference on Multimedia.2017:1856-1864.
[43]LI C L,LU A,ZHENG A H,et al.Multi-adapter RGBT tra-cking[C]//2019 IEEE/CVF International Conference on Computer Vision Workshop(ICCVW).IEEE,2019:2262-2270.
[44]HARE S,GOLODETZ S,SAFFARI A,et al.Struck:Structured output tracking with kernels[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2015,38(10):2096-2109.
[45]VALMADRE J,BERTINETTO L,HENRIQUES J,et al.End-to-end representation learning for correlation filter based tra-cking[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:2805-2813.
[46]HENRIQUES J F,CASEIRO R,MARTINS P,et al.High-speed tracking with kernelized correlation filters[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2014,37(3):583-596.
[47]WU Y,BLASCH E,CHEN G,et al.Multiple source data fusion via sparse representation for robust visual tracking[C]//14th International Conference on Information Fusion.IEEE,2011:1-8.
[48]LIU H P,SUN F C.Fusion tracking in color and infrared images using joint sparse representation[J].Science China Information Sciences,2012,55(3):590-599.
[1] GUO Yan-qing, LI Yu-hang, WANG Wan-wan, FU Hai-yan, WU Ming-kan, LI Yi. FL-GRM:Gamma Regression Algorithm Based on Federated Learning [J]. Computer Science, 2022, 49(12): 66-73.
[2] JIN Yu-jie, CHU Xu, WANG Ya-sha, ZHAO Jun-feng. Variational Domain Adaptation Driven Semantic Segmentation of Urban Scenes [J]. Computer Science, 2022, 49(11): 126-133.
[3] ZHANG Yu-xin, CHEN Yi-qiang. Driver Distraction Detection Based on Multi-scale Feature Fusion Network [J]. Computer Science, 2022, 49(11): 170-178.
[4] SONG Mei-qi, FU Xiang-ling, YAN Chen-wei, WU Wei-qiang, REN Yun. Prediction Model of Enterprise Resilience Based on Bi-directional Long Short-term Memory Network [J]. Computer Science, 2022, 49(11): 197-205.
[5] ZHONG Kun-hua, CHEN Yu-wen, QIN Xiao-lin. Sub-BN-Merge Based Bayesian Network Structure Learning Algorithm [J]. Computer Science, 2022, 49(11A): 210800172-7.
[6] REN Shuang-yan, GUO Wei, FAN Chang-qi, WANG Zhe, WU Song-yang. Multi-view Distance Metric Learning with Inter-class and Intra-class Density [J]. Computer Science, 2022, 49(11A): 211000131-6.
[7] XU Hui, WANG Zhong-qing, LI Shou-shan, ZHANG Min. Personalized Dialogue Generation Integrating Sentimental Information [J]. Computer Science, 2022, 49(11A): 211100019-6.
[8] CEN Jian-ming, FENG Quan-xi, ZHANG Li-li, TONG Rui-chao. Empirical Study on the Forecast of Large Stock Dividends of Listed Companies Based on DE-lightGBM [J]. Computer Science, 2022, 49(11A): 211000017-7.
[9] SUN Kai-wei, GUO Hao, ZENG Ya-yuan, FANG Yang, LIU Qi-lie. Multi-target Regression Method Based on Hypernetwork [J]. Computer Science, 2022, 49(11A): 211000205-9.
[10] WU Xiao-wen, ZHENG Qiao-xian, XU Xin-qiang. Improved Ant Colony Algorithm for Solving Multi-objective Unilateral Assembly Line Balancing Problem [J]. Computer Science, 2022, 49(11A): 210900165-5.
[11] WANG Mao-guang, JI Hao-yue, WANG Tian-ming. Study on Risk Control Model of Selective Ensemble Algorithm Based on Hierarchical Clustering and Simulated Annealing [J]. Computer Science, 2022, 49(11A): 210800105-7.
[12] DAI Xiao-lu, WANG Ting-hua, ZHOU Hui-ying. Fuzzy Multiple Kernel Support Vector Machine Based on Weighted Mahalanobis Distance [J]. Computer Science, 2022, 49(11A): 210800216-5.
[13] XU Wei-hua, ZHANG Jun-jie, CHEN Xiu-wei. Distribution Reduction in Fuzzy Order Decision Data Sets with Attention Degree [J]. Computer Science, 2022, 49(11A): 210700191-5.
[14] WANG Zhi-qiang, ZHENG Ting-ting, SUN Xin, LI Qing. Attribute Reduction Algorithm Based on a New q-rung orthopair Fuzzy Cross Entropy [J]. Computer Science, 2022, 49(11A): 211200142-6.
[15] RAN Hong, HOU Ting, HE Long-yu, QIN Ke-yun. Fuzzy Rough Sets Model Based on Fuzzy Neighborhood Systems [J]. Computer Science, 2022, 49(11A): 211100224-5.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!