计算机科学 ›› 2023, Vol. 50 ›› Issue (12): 279-284.doi: 10.11896/jsjkx.221000245
许华杰1,2,3,4, 肖毅烽1
XU Huajie1,2,3,4, XIAO Yifeng1
摘要: 基于一致性正则化的方法在半监督语义分割任务中展现出了较好的性能,这类方法通常涉及两个角色:一个显式或隐式的教师网络和一个学生网络。其中学生网络通过最小化两个网络对不同扰动样本预测结果之间的一致性损失实现训练。但是来自单个教师网络的不可靠预测可能会导致学生网络学习到错误的信息。通过将平均教师模型MT的单教师网络扩展为多教师网络,提出了多平均教师网络(Multiple Mean Teacher Network,MMTNet)模型,使学生网络从多个教师网络的平均预测结果进行学习,有效降低单个教师网络预测错误的影响。此外,MMTNet通过对无标签数据进行强、弱数据增强的方式对无标签数据进行数据扰动,增加了无标签数据的多样性,在一定程度上缓解了学生网络和教师网络之间存在的耦合问题,避免了学生网络对教师网络的过度拟合,从而进一步降低了教师网络进行伪标签预测错误时所产生的影响。在PASCAL VOC 2012扩充数据集上的实验结果表明,所提出的多平均教师网络MMTNet模型可获得比其他目前主流的半监督语义分割方法更高的平均交并比,且实际分割效果更优。
中图分类号:
[1]CHEN L C,ZHU Y,PAPANDREOU G,et al.Encoder-decoder with atrous separable convolution for semantic image segmentation[C]//European Conference on Computer Vision(ECCV).2018:801-818. [2]TIAN X,WANG L,DING Q.Review of Image Semantic Seg-mentation Based on Deep Learning[J].Journal of Software,2019,30(2):440-468. [3]BEARMAN A,RUSSAKOVSKY O,FERRARI V,et al.What'sthe point:Semantic segmentation with point supervision[C]//European Conference on Computer Vision.Cham:Springer,2016:549-565. [4]FRENCH G,LAINE S,AILA T,et al.Semi-supervised semantic segmentation needs strong,varied perturbations[C]//British Machine Vision Conference.2020:1-7. [5]CHEN X,YUAN Y,ZENG G,et al.Semi-supervised semantic segmentation with cross pseudo supervision[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2021:2613-2622. [6]OUALI Y,HUDELOT C,TAMI M.Semi-supervised semantic segmentation with cross-consistency training[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2020:12674-12684. [7]TARVAINEN A,VALPOLA H.Mean teachers are better role models:Weight-averaged consistency targets improve semi-supervised deep learning results[J].Advances in Neural Information Processing Systems,2017,30:2-7. [8]REN Z,YEH R,SCHWING A.Not all unlabeled data areequal:Learning to weight data in semi-supervised learning[J].Advances in Neural Information Processing Systems,2020,33:21786-21797. [9]YANG L,ZHUO W,QI L,et al.St++:Make self-trainingwork better for semi-supervised semantic segmentation[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:4268-4277. [10]SOHN K,BERTHELOT D,CARLINI N,et al.Fixmatch:Simplifying semi-supervised learning with consistency and confidence[J].Advances in Neural Information Processing Systems,2020,33:596-608. [11]XU Y,SHANG L,YE J,et al.Dash:Semi-supervised learning with dynamic thresholding[C]//International Conference on Machine Learning.PMLR,2021:11525-11536. [12]LIU S P,HONG J M,LIANG J P,et al.Medical Image Segmentation Using Semi-supervised Conditional Generative Adversa-rial Nets[J].Journal of Software,2020,31(8):2588-2602. [13]LI Z X,ZHANG J,WU J L,et al.Semi-supervised adversarial learning based semantic image segmentation[J].Journal ofImage and Graphics,2022,27(7):2157-2170. [14]GUAN D,HUANG J,XIAO A,et al.Unbiased subclass regularization for semi-supervised semantic segmentation[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:9968-9978. [15]KWON D,KWAK S.Semi-supervised semantic segmentationwith error localization network[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:9957-9967. [16]KE R,AVILES-RIVERO A I,PANDEY S,et al.A three-stage self-training framework for semi-supervised semantic segmentation[J].IEEE Transactions on Image Processing,2022,31:1805-1815. [17]WANG Y,WANG H,SHEN Y,et al.Semi-supervised semantic segmentation using unreliable pseudo-labels[C]//IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:4248-4257. [18]ZOU Y,ZHANG Z,ZHANG H,et al.PseudoSeg:DesigningPseudo Labels for Semantic Segmentation[C]//International Conference on Learning Representations.2021:1-8. [19]EVERINGHAM M,ESLAMI S A,VAN GOOL L,et al.Thepascal visual object classes challenge:A retrospective[J].International Journal of Computer Vision,2015,111(1):98-136. [20]YUN S,HAN D,OH S J,et al.Cutmix:Regularization strategy to train strong classifiers with localizable features[C]//IEEE/CVF International Conference on Computer Vision.2019:6023-6032. |
|