Computer Science ›› 2025, Vol. 52 ›› Issue (12): 150-157.doi: 10.11896/jsjkx.241200021

• Computer Graphics & Multimedia • Previous Articles     Next Articles

ETF-YOLO11n:Object Detection Method Based on Multi-scale Feature Fusion for TrafficImages

XIA Shufang, YIN Haonan, QU Zhong   

  1. School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
  • Received:2024-12-02 Revised:2025-03-30 Online:2025-12-15 Published:2025-12-09
  • About author:XIA Shufang,born in 1980,Ph.D.Her main research interests include compu-ter vision,machine learning and artificial intelligence.
    QU Zhong,born in 1972,Ph.D,professor.His main research interests include computer vision,machine learning and artificial intelligence.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China(62576058,62571077),Scientific and Technological Research Program of Chongqing Municipal Education Commission(KJZD-M202300604) and Natural Science Foundation of Chongqing,China(2023NSCQ-MSX1781).

Abstract: Deep learning algorithms have made significant progress in the field of computer vision in recent years,but the accuracy of object detection in complex traffic scenes is still unsatisfactory due to the small size of traffic objects,inconspicuous feature,and susceptibility to interference.To address this problem,this paper improves the state-of-the-art YOLO11 and designs the ETF-YOLO11n based on multi-scale feature fusion.Firstly,it designs TFF,which effectively fuses the feature information of different sizes extracted from the backbone.Secondly,it designs HDCFE,effectively integrates the features extracted from different receptive fields and reduces the interference on the detection effect of the model due to occlusion and overlapping.Finally,the proposed GeoCIoU is used to replace CIoU,and the model can provide more accurate feedback on the matching of the predicted box and the ground-truth box through the two different penalization terms.The ETF-YOLO11n achieves an AP of 65.6% and mAP@0.5 of 90.7% on KITTI dataset,which is improved by 2.4 percentage points and 1.2 percentage points.In addition,ETF-YOLO11n achieves 42.5% AP and 59.8% mAP@0.5 on COCO-Traffic,and EFT-YOLOv8n achieves 66.9% AP and 91.5% mAP@0.5 on KITTI dataset.The results show that the proposed methods significantly improve the performance and have good ge-neralization ability to different models and datasets,achieve a good balance between the accuracy and parameters.The source code has been opened.

Key words: Object detection, Multi-feature fusion, Inter over Union, Feature enhancement, Complex traffic scenarios

CLC Number: 

  • TP391.41
[1]CHEN H,WAN W W,MATSUSHIT A,et al.AutomaticallyPrepare Training Data for YOLO Using Robotic In-Hand Observation and Synthesis [J].IEEE Transactions on Automation Science and Engineering,2024,21(3):4876-4982.
[2]LIU W,ANGUELOV D,ERHAN D,et al.SSD:Single ShotMultibox Detector [C]//European Conference on Computer Vision.Springer,2016:21-27.
[3]JOCHER G.YOLOv5[EB/OL].(2020-06-28) [2024-11-17].https://github.com/ultralytics/yolov5.
[4]LI C Y,LI LL,JIANG H L,et al.YOLOv6:A Single-Stage Object Detection Framework for Industrial Applications[J].arXiv:2209.02976,2022.
[5]ZHENG G,LIU S T,WANG F,et al.YOLOX:Exceeding YOLO Series in 2021[J].arXiv:2107.08430,2021.
[6]WANG C Y,BOCHKOVSKIY A,LIAO H Y M.YOLOv7:Trainable Bag-of-freebies Sets New State-of-the-art for Real-time Object Detectors[C]//IEEE Conference on Computer Vision and Pattern Recognition.2023:7464-7475.
[7]JOCHER G.YOLOv8[EB/OL].(2023-01-10) [2024-11-17].https://github.com/ultralytics/yolov8.
[8]WANG C Y,YEH I H,LIAO H Y,et al.Yolov9:LearningWhat You Want to Learn Using Programmable Gradient Information[C]//European Conference on Computer Vision.Sprin-ger,2024:1-21.
[9]JOCHER G.YOLO11[EB/OL].(2024-09-30) [2024-11-17].https://GitHub-ultralytics/ultralytics:Ultralytics YOLO11.
[10]GIRSHICK R,REN S,HE K,et al.Faster R-CNN:TowardsReal-time Object Detection with Region Proposal Networks [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2017,39(6):1137-1149.
[11]CARION N,MASSA F,SYNNAEVE G,et al.End-to-end Object Detection with Transformers[C]//European Conference on Computer Vision.Springer,2020:213-219.
[12]Government of the People’s Republic of China.MotorVehicles Nationwide Reached 440 Million in the First Half of 2024 [EB/OL].(2024-07-09) [2024-11-17].https://www.gov.cn/lianbo/bumen/202407/content_6961935.htm.
[13]GEIGER A,LENZ P,STILLER C,et al.Vision Meets Robo-tics:The Kitti Dataset[J].The International Journal of Robotics Research,2013,32(11):1231-1237.
[14]LIN T Y,DOLLAR P,PIOTR G,et al.FeaturePyramid Networks for Object Detection [C]//IEEE Conference on Compu-ter Vision and Pattern Recognition.IEEE,2017:936-944.
[15]LIU S,QI L,QIN H F,et al.Path Aggregation Network for Instance Segmentation [C]//IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2018:8759-8768.
[16]LAN L X,CHI M Y.Remote Sensing Change Detection Based on Feature Fusion and Attention Network[J].Computer Science,2022,49(6):193-198.
[17]LU H T,FANG M Y,QIU Y X,et al.An Anchor-Free Defect Detector for Complex Background Based on Pixelwise Adaptive Multiscale Feature Fusion[J].IEEE Transactions on Instrumentation and Measurement,2023,72:1-12.
[18]YU J H,JIANG Y N,WANG Z Y.et al.Unitbox:An advanced object detection network [C]//International Conference on Multimedia.ACM,2016:516-520.
[19]REZATOFIGHI H,HAMID T,NATHAN G,et al.Generalized Intersection over Union:A Metric and a Loss for Bounding Box Regression[C]//IEEE International Conference on Computer Vision.IEEE,2019:658-666.
[20]ZHENG Z H,WANG P,LIU W,et al.Distance-IoU loss:Faster and Better Learning for Bounding Box Regression [C]//AAAI Conference on Artificial Intelligence.AAAI,2020:12993-13000.
[21]GEVORGYAN Z.SIoU Loss:More Powerful Learning forBounding Box Regression [J].arXiv:2205.12740,2022.
[22]LUO X,CAI Z,SHAO B,et al.Unified-IoU:For High-Quality Object Detection [J].arXiv:2408.06636,2024.
[23]HU J,SHEN L,SUN G.Squeeze-and-excitation Networks[C]//IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2018:7132-7141.
[24]WOO S P,LEE J C,KWEON J Y.CBAM:Convolutional Block Attention Module [C]//European Conference on Computer Vision.Springer,2018:3-19.
[25]REN X X,LI M,LI Z H,et al.Curiosity-driven Attention forAno-maly Road Obstacles Segmentation in Autonomous Driving [J].IEEE Transactions on Intelligent Vehicles,2022,8(3):2233-2243.
[26]VASWANI A,SHAZEER N,PARMAR N,et al.Attention is All You Need[J].arXiv:1706.03762,2017.
[27]ANTHONY F,DANIEL G,KYROLLOS Y Y,et al.LookHere:Vision Transformers with Directed Attention Generalize and Extrapolate[J].arXiv:2405.13958,2024.
[28]GAO L Y,QU Z,WANG S Y,et al.A Lightweight Neural Network Model of Feature Pyramid and Attention Mechanism for Traffic Object Detection[J].IEEE Transactions on Intelligent Vehicles,2024,9(2):3422-3435.
[29]WANG S Y,QU Z,GAO L Y,et al.Multi-spatial Pyramid Feature and Optimizing Focal Loss Function for Object Detection[J].IEEE Transactions on Intelligent Vehicles,2023,9(1):1054-1065.
[30]LIN T S,GOYAL P,GIRSHICK R,et al.Focal Loss for Dense Object Detection [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2017,99:2999-3007.
[1] SHEN Tao, ZHANG Xiuzai, XU Dai. Improved RT-DETR Algorithm for Small Object Detection in Remote Sensing Images [J]. Computer Science, 2025, 52(8): 214-221.
[2] LIU Chengzhuang, ZHAI Sulan, LIU Haiqing, WANG Kunpeng. Weakly-aligned RGBT Salient Object Detection Based on Multi-modal Feature Alignment [J]. Computer Science, 2025, 52(7): 142-150.
[3] XU Yongwei, REN Haopan, WANG Pengfei. Object Detection Algorithm Based on YOLOv8 Enhancement and Its Application Norms [J]. Computer Science, 2025, 52(7): 189-200.
[4] WANG Rui, TANG Zhanjun. Multi-feature Fusion and Ensemble Learning-based Wind Turbine Blade Defect Detection Method [J]. Computer Science, 2025, 52(6A): 240900138-8.
[5] DING Xuxing, ZHOU Xueding, QIAN Qiang, REN Yueyue, FENG Youhong. High-precision and Real-time Detection Algorithm for Photovoltaic Glass Edge Defects Based onFeature Reuse and Cheap Operation [J]. Computer Science, 2025, 52(6A): 240400146-10.
[6] HUANG Hong, SU Han, MIN Peng. Small Target Detection Algorithm in UAV Images Integrating Multi-scale Features [J]. Computer Science, 2025, 52(6A): 240700097-5.
[7] LE Lingzhi, ZHAI Jiangtao, YU Ming, SUN Tongqing. Object Detection-based Method for Guiding Passenger Flow in Boarding and Deparking Areas ofRail Transit [J]. Computer Science, 2025, 52(6A): 240400192-9.
[8] ZHANG Dabin, WU Qin, ZHOU Haojie. Oriented Object Detection Based on Multi-scale Perceptual Enhancement [J]. Computer Science, 2025, 52(6): 247-255.
[9] LI Xiaolan, MA Yong. Study on Lightweight Flame Detection Algorithm with Progressive Adaptive Feature Fusion [J]. Computer Science, 2025, 52(4): 64-73.
[10] ZHOU Yi, MAO Kuanmin. Research on Individual Identification of Cattle Based on YOLO-Unet Combined Network [J]. Computer Science, 2025, 52(4): 194-201.
[11] LI Zongmin, RONG Guangcai, BAI Yun, XU Chang , XIAN Shiyang. 3D Object Detection with Dynamic Weight Graph Convolution [J]. Computer Science, 2025, 52(3): 104-111.
[12] CHENG Qinghua, JIAN Haifang, ZHENG Shuaikang, GUO Huimin, LI Yuehao. Illumination-aware Infrared/Visible Fusion for Object Detection [J]. Computer Science, 2025, 52(2): 173-182.
[13] GAO Yuli, WANG Baohui. DEFM-YOLOv8-based Detection Algorithm for High-speed Rail Contact Network Wire State [J]. Computer Science, 2025, 52(11A): 241000155-9.
[14] WANG Hongqiang, ZHAO Hui, JIA Zhenhong. Cotton Disease Detection Based on Feature Enhancement and Group Mix Attention [J]. Computer Science, 2025, 52(11A): 250200043-7.
[15] DUAN Pengsong, GAO Yang, ZHANG Dalong, CAO Yangjie, ZHAO Jie. C2P-YOLO:A Lightweight Crack Detection Algorithm for Wind Turbine Towers [J]. Computer Science, 2025, 52(11A): 250100126-6.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!