计算机科学 ›› 2024, Vol. 51 ›› Issue (11A): 231000071-9.doi: 10.11896/jsjkx.231000071

• 图像处理&多媒体技术 • 上一篇    下一篇

基于非特定类别图像前景主体分割的深度学习算法研究

陈祥龙, 李海军   

  1. 三亚学院信息与智能工程学院 海南 三亚 572022
    三亚学院陈国良院士团队创新中心 海南 三亚 572022
  • 出版日期:2024-11-16 发布日期:2024-11-13
  • 通讯作者: 李海军(haijunli1968@163.com)
  • 作者简介:(1482103902@qq.com)

Study on Deep Learning Algorithm for Foreground Subject Segmentation of Non-specific CategoryImages

CHEN Xianglong, LI Haijun   

  1. School of Information and Intelligent Engineering,University of Sanya,Sanya,Hainan 572022,China
    Academician Guoliang Chen Team Innovation Center,University of Sanya,Sanya,Hainan 572022,China
  • Online:2024-11-16 Published:2024-11-13
  • About author:CHEN Xianglong,born in 2001,postgraduate,is a member of CCF(No.Q5078G).His main research interests include computer vision and data mining.
    LI Haijun,born in 1968,Ph.D,associate professor,master's supervisor,is a member of CCF(No.F7747M).His main research interests include compu-ter vision and data mining.

摘要: 通过在Mobile-Unet网络的基础上加入SENet通道注意力机制来改进图像前景主体分割算法。改进后的算法引入深度可分离卷积来减小模型参数量,同时利用跳跃连接和多尺度特征融合来提高模型的分割精度。在训练过程中,采用了带空洞卷积的空间金字塔池化模块来增加感受野,提高模型对于大尺度物体的识别能力。实验结果表明,改进后的算法在PASCAL VOC2012数据集上达到了96%的MIOU(Modular Input/Output Unit)分割精度,准确率达到了0.971,优于现有的多种图像分割算法,例如FCN全卷积神经网络算法。在速度方面,模型对于每张图片的处理时间为1.7~2.5 s,改进后的算法相对于传统的全卷积神经网络具有更快的推理速度,适合于在移动设备上实现实时图像分割。通过对比实验,比较了改进前和改进后的Mobile-Unet模型以及FCN模型对于明亮条件下和昏暗条件下图像前景主体分割的效果,并得出了改进后的Mobile-Unet模型具有最好效果的结论。最终进行算法的部署,设计了GUI可视化操作界面,并生成.exe可执行文件。

关键词: 主体分割, 神经网络, 感受野, 参数量, 分割精度

Abstract: By incorporating SENet channel attention mechanism on the basis of Mobile Unet network,the image foreground subject se-gmentation algorithm is improved.The algorithm introduces deep separable convolution to reduce the number of model parameters,while utilizing skip connections and multi-scale feature fusion to improve the segmentation accuracy of the model.Du-ring the training process,a spatial pyramid pooling module with hollow convolution is used to increase the receptive field and improve the model's recognition ability for large-scale objects.Experimental results show that the improved algorithm achieves 96% MIOU(Modular Input/Output Un-it) segmentation accuracy on the PASCAL VOC2012 dataset,with an accuracy rate of 0.971,which is superior to various existing image segmentation algorithms,such as the FCN fully convolutional neural network algorithm.In terms of speed,the processing time of the model for each image is between 1.7 s and 2.5 s.The improved algorithm has a faster inference speed compared to traditional fully convolutional neural networks,making it suitable for real-time image segmentation on mobile devices.Through comparative experiments,the effectiveness of the Mobile Unet models before and after the improvement,as well as the FCN model,in foreground subject segmentation of images under bright and dim conditions is compared,and the conclusion is drawn that the improved Mobile Unet model has the best performance.Finally,the algorithm is deployed,a GUI visualization operation interface is designed,and an.exe executable file is generated.

Key words: Subject segmentation, Neural network, Receptive field, Parameter quantity, Segmentation accuracy

中图分类号: 

  • TP183
[1]LIU Y,YANG S.Shadow segmentation of indoor moving objects based on improved UNet network [J].Computer System Application,2022,31(12):412-419.
[2]MENG M Z,LI L,HE G Y,et al.Preliminary study on the deg-radation effect of MobileNetV2 on four types of breast X-ray BI-RADS lesions [J].Journal of Clinical Radiology,2022,41(10):1868-1873.
[3]GUOFENG M.A Small Target Detection Method Based on the Improved FCN Model[J].Advances in Multimedia,2022,2022.
[4]YIMING L,ZHENGLE W,RUJIA W,et al.Flooding-basedMobileNet toidentify cucumber diseases from leaf images in natural scenes[J].Computers and Electronics in Agriculture,2023,213.
[5]CHEN L C,PAPANDREOV G,KOKKINOS I,et al.DeepLab:Semantic Image Segmentation with Deep Convolutional Nets,Atrous Convolution,and Fully Connected CRFs.[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2018,40(4):834-838.
[6]WU D,ZHAO J,WANG Z.AM-PSPNet:Pyramid Scene Parsing Network Based on Attentional Mechanism for Image Semantic Segmentation[C]//ICPCSEE Steering Committee.Abstracts of the 8th International Conference of Pioneering Computer Scientists,Engineers and Educators(ICPCSEE 2022).Part I.Springer,2022:444.
[7]ZHANG X,YAO Q A,ZHAO J,et al.Overview of Full Convo-lutional Neural Network Image Semantic Segmentation Methods [J].Computer Engineering and Applications,2022,58(8):45-57.
[8]HUANG P,ZHENG Q,LIANG O.Overview of Image Segmentation Methods [J].Journal of Wuhan University(Science Edition),2020,66(6):519-531.
[9]HUANG S P,LIU H N,ZHOU K S,et al.Zebra crossing segmentation based on improved Unet [J].Intelligent Computer and Applications,2020,10(11):61-64,69.
[10]SUN L X.Coal gangue image recognition based on convolutional neural network [J].Computer Knowledge and Technology,2020,16(21):16-18,22.
[11]LI Y Q.Research and Implementation of Image Saliency Detection Algorithm Based on Attention Mechanism and U-Net [D].Beijing:Beijing Jiaotong University,2020.
[12]WANG P,GAO C,ZHU L,et al.Ischemic stroke lesion segmentation algorithm based on 3D deep residual network and cascaded U-Net [J].Computer Application,2019,39(11):3274-3279.
[13]DING R J,GAO F F,XING L.An Intelligent Routing Strategy for the Internet of Things Based on Deep Reinforcement Learning [J].Journal of the Internet of Things,2019,3(2):56-63.
[14]TIAN X,WANG L,DING Q.Overview of Image Semantic Segmentation Methods Based on Deep Learning [J].Journal of Software Science,2019,30(2):440-468.
[15]ZHANG M Y.Research on Image Segmentation Based on Deep Learning [D].Jilin:Jilin University,2017.
[16]CHEN H X.Image Semantic Segmentation Based on Convolutional Neural Networks [D].Hangzhou:Zhejiang University,2016.
[17]LIU S T,YIN F L.Image segmentation methods based on graph cuts and their new progress [J].Journal of Automation,2012,38(6):911-922.
[18]HE J,GE H,WANG Y F.Overview of Image Segmentation Algorithms [J].Computer Engineering and Science,2009,31(12):58-61.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!