Computer Science ›› 2022, Vol. 49 ›› Issue (10): 159-168.doi: 10.11896/jsjkx.210800050

• Computer Graphics& Multimedia • Previous Articles     Next Articles

Robust Hash Learning Method Based on Dual-teacher Self-supervised Distillation

MIAO Zhuang, WANG Ya-peng, LI Yang, WANG Jia-bao, ZHANG Rui, ZHAO Xin-xin   

  1. Command and Control Engineering College,Army Engineering University of PLA,Nanjing 210007,China
  • Received:2021-08-05 Revised:2021-12-08 Online:2022-10-15 Published:2022-10-13
  • About author:MIAO Zhuang,born in 1976,associate professor,is a member of China Computer Federation.His main research interests include artificial intelligence,pattern recognition and computer vision.
    LI Yang,born in 1984,associate professor,is a member of China Computer Federation.His main research interests include computer vision,deep learning and image processing.
  • Supported by:
    National Natural Science Foundation of China(61806220)and National Key Research and Development Program of China(2017YFC0821905).

Abstract: In order to improve the performance of unsupervised hash learning and achieve robust hashing image retrieval,this paper proposes a novel robust hash learning method based on dual-teacher self-supervised distillation.Specifically,the proposed method contains two stages:a self-supervised dual-teacher learning stage and a robust hash learning stage.In the first stage,a modified cluster algorithm is designed to effectively improve the accuracy of hard pseudo labels.Then,we fine-tune the teacher networks by hard pseudo labels to get the initial soft pseudo labels.In the second stage,we filter the initial soft pseudo labels by our soft pseudo label denoising method,which combines a hybrid denoising strategy and a dual-teacher denoising strategy.Then,we train the student network with the denoised soft pseudo labels by knowledge distillation,so that robust hash codes for label-free images are obtained.Extensive experiments on CIFAR-10,FLICKR25K and EuroSAT datasets show that the proposed robust hash learning method outperforms the state-of-the-art methods.In detail,the MAP of our method is 18.6% higher than that of the TBH method on CIFAR-10,2.4% higher than that of the DistillHash method on FLICKR25K,and 18.5% higher than that of the ETE-GAN method on EuroSAT.

Key words: Hash learning, Self-supervised learning, Knowledge distillation, Image retrieval, Noisy labels

CLC Number: 

  • TP391
[1]WANG J,KUMAR S,CHANG S F.Semi-supervised hashing for large-scale search[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2012,34(12):2393-2406.
[2]SHARIF U,MEHMOOD Z,MAHMOOD T,et al.Scene analysis and search using local features and support vector machine for effective content-based image retrieval[J].Artificial Intelligence Review,2019,52(2):901-925.
[3]PENG Y,ZHANG J,YE Z.Deep reinforcement learning forimage hashing[J].IEEE Transactions on Multimedia,2019,22(8):2061-2073.
[4]ZHANG B,QIAN J.Autoencoder-based unsupervised clustering and hashing[J].Applied Intelligence,2021,51(1):493-505.
[5]GOODFELLOW I J,POUGET-ABADIE J,MIRZA M,et al.Generative adversarial networks[C]//Proceedings of Advances in Neural Information Processing Systems.2014:2672-2680.
[6]SHEN Y,QIN J,CHEN J,et al.Auto-encoding twin-bottleneck hashing[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2020:2815-2824.
[7]CHEN X,LU C.An end-to-end adversarial hashing method for unsupervised multispectral remote sensing image retrieval[C]//Proceedings of the IEEE International Conference on Image Processing.IEEE,2020:1536-1540.
[8]SONG J,HE T,GAO L,et al.Binary generative adversarial net-works for image retrieval[C]//Proceedings of the AAAI Conference on Artificial Intelligence.AAAI,2018:394-401.
[9]YANG E,LIU T,DENG C,et al.Distillhash:unsupervised deep hashing by distilling data pairs[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2019:2946-2955.
[10]YANG E,DENG C,LIU T,et al.Semantic structure-based unsupervised deep hashing[C]//Proceedings of the International Joint Conference on Artificial Intelligence.2018:1064-1070.
[11]DIZAJI G,ZHENG F,SADOUGHI N,et al.Unsupervised deep generative adversarial hashing network[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2018:3664-3673.
[12]LIN K,LU J,CHEN C S,et al.Learning compact binary de-scriptors with unsupervised deep neural networks[C]//Procee-dings of the IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2016:1183-1192.
[13]DAI B,GUO R,KUMAR S,et al.Stochastic generative hashing[C]//Proceedings of the International Conference on Machine Learning.PMLR,2017:913-922.
[14]LI Y,WANG Y,MIAO Z,et al.Contrastive self-supervised hashing with dual pseudo agreement[J].IEEE Access,2020,8:165034-165043.
[15]HUANG J,DONG Q,GONG S,et al.Unsupervised deep lear-ning by neighbourhood discovery[C]//Proceedings of the International Conference on Machine Learning.PMLR,2019,97:2849-2858.
[16]CHEN T,KORNBLITH S,NOROUZI M,et al.A simpleframework for contrastive learning of visual representations[C]//Proceedings of the International Conference on Machine Learning.PMLR,2020,119:1597-1607.
[17]HE K,FAN H,WU Y,et al.Momentum contrast for unsupervised visual representation learning[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2020:9726-9735.
[18]BACHMAN P,HJELM R,BUCHWALTER W.Learning representations by maximizing mutual information across views[C]//Proceedings of Advances in Neural Information Proces-sing Systems.2019:15509-15519.
[19]CARON M,BOJANOWSKI P,JOULIN A,et al.Deep clustering for unsupervised learning of visual features[C]//Procee-dings of the European Conference on Computer Vision.Berlin:Springer,2018:139-156.
[20]ASANO Y,RUPPRECHT C,VEDALDI A.Self-labelling viasimultaneous clustering and representation learning[C]//Proceedings of the International Conference on Learning Representations,OpenReview,2020.
[21]CARON M,MISRA I,MAIRAL J,et al.Unsupervised learning of visual features by contrasting cluster assignments[C]//Proceedings of Advances in Neural Information Processing Systems.2020.
[22]HAN B,YAO Q,LIU T,et al.A survey of label-noise representation learning:past,present and future[EB/OL].(2021-02-20) [2021-04-27].https://arxiv.org/abs/2011.04406.
[23]REED S,LEE H,ANGUELOV D,et al.Training deep neuralnetworks on noisy labels with bootstrapping[C]//Proceedings of the International Conference on Learning Representations.2015.
[24]SHU J,XIE Q,YI L,et al.Meta-weight-net:Learning an expli-cit mapping for sample weighting[C]//Proceedings of Advances in Neural Information Processing Systems.2019:1919-1930.
[25]LI Y,YANG J,SONG Y,et al.Learning from noisy labels with distillation[C]//Proceedings of the International Conference on Computer Vision.IEEE,2017:1928-1936.
[26]HUANG Z,ZOU Y,BHAGAVATULA V,et al.Comprehen-sive attention self-distillation for weakly-supervised object detection[C]//Proceedings of Advances in Neural Information Processing Systems.2020.
[27]HINTON G,VINYALS O,DEAN J.Distilling the knowledge in a neural network[J].Computer Science,2015,14(7):38-39.
[28]NORTHCUTT C,JIANG L,CHUANG I.Confident Learning:Estimating Uncertainty in Dataset Labels[J].Journal of Artificial Intelligence Research,2021,70:1373-1411.
[29]WEISS Y,TORRALBA A,FERGUS R.Spectral hashing[C]//Proceedings of Advances in Neural Information Processing Systems,Curran Associates,Inc..2008:1753-1760.
[30]ANDONI A,INDYK P.Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions[J].Communications of the ACM,2008,51(1):117-122.
[31]WANG X,ZHANG L,JING F,et al.AnnoSearch:Image auto-annotation by search[C]//Proceedings of the 2006 IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2006:1483-1490.
[32]HEO J,LEE Y,HE J,et al.Spherical hashing[C]//Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2012:2957-2964.
[33]JIN Z,LI C,LIN Y,et al.Density sensitive hashing[J].IEEE Transactions on Cybernetics,2014,44(8):1362-1371.
[34]GONG Y,LAZEBNIK S,GORDO A,et al.Iterative quantiza-tion:A procrustean approach to learning binary codes for large-scale image retrieval[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2013,35(12):2916-2929.
[35]WEI Y,TIAN D,SHI J,et al.Optimizing non-differentiablemetrics for hashing[J].IEEE Access,2021,9:14351-14357.
[36]RAGINSKY M,LAZEBNIK S.Locality-sensitive binary codesfrom shift-invariant kernels[C]//Proceedings of Advances in Neural Information Processing Systems,Curran Associates,Inc..2009:1509-1517.
[37]XIA Y,HE K,KOHLI P,et al.Sparse projections for high-dimensional binary codes[C]//Proceedings of the IEEE Confe-rence on Computer Vision and Pattern Recognition.IEEE,2015:3332-3339.
[1] CAO Xiao-wen, LIANG Mei-yu, LU Kang-kang. Fine-grained Semantic Reasoning Based Cross-media Dual-way Adversarial Hashing Learning Model [J]. Computer Science, 2022, 49(9): 123-131.
[2] ZHU Cheng-zhang, HUANG Jia-er, XIAO Ya-long, WANG Han, ZOU Bei-ji. Deep Hash Retrieval Algorithm for Medical Images Based on Attention Mechanism [J]. Computer Science, 2022, 49(8): 113-119.
[3] DU Hang-yuan, LI Duo, WANG Wen-jian. Method for Abnormal Users Detection Oriented to E-commerce Network [J]. Computer Science, 2022, 49(7): 170-178.
[4] CHU Yu-chun, GONG Hang, Wang Xue-fang, LIU Pei-shun. Study on Knowledge Distillation of Target Detection Algorithm Based on YOLOv4 [J]. Computer Science, 2022, 49(6A): 337-344.
[5] CHENG Xiang-ming, DENG Chun-hua. Compression Algorithm of Face Recognition Model Based on Unlabeled Knowledge Distillation [J]. Computer Science, 2022, 49(6): 245-253.
[6] HUANG Zhong-hao, YANG Xing-yao, YU Jiong, GUO Liang, LI Xiang. Mutual Learning Knowledge Distillation Based on Multi-stage Multi-generative Adversarial Network [J]. Computer Science, 2022, 49(10): 169-175.
[7] TIAN Song-wang, LIN Su-zhen, YANG Bo. Multi-band Image Self-supervised Fusion Method Based on Multi-discriminator [J]. Computer Science, 2021, 48(8): 185-190.
[8] WANG Chun-jing, LIU Li, TAN Yan-yan, ZHANG Hua-xiang. Image Retrieval Method Based on Fuzzy Color Features and Fuzzy Smiliarity [J]. Computer Science, 2021, 48(8): 191-199.
[9] WANG Xiao-fei, ZHOU Chao, LIU Li-gang. Lightweight Image Retrieval System Based on Feature Clustering [J]. Computer Science, 2021, 48(2): 148-152.
[10] YU Liang, WEI Yong-feng, LUO Guo-liang, WU Chang-xing. Knowledge Distillation Based Implicit Discourse Relation Recognition [J]. Computer Science, 2021, 48(11): 319-326.
[11] XU Xing, SUN Jia-liang, WANG Zheng, YANG Yang. Feature Transformation for Defending Adversarial Attack on Image Retrieval [J]. Computer Science, 2021, 48(10): 258-265.
[12] WANG Run-zheng, GAO Jian, HUANG Shu-hua, TONG Xin. Malicious Code Family Detection Method Based on Knowledge Distillation [J]. Computer Science, 2021, 48(1): 280-286.
[13] JIAO Yang, YANG Chuan-ying, SHI Bao. Relevance Feedback Method Based on SVM in Shoeprint Images Retrieval [J]. Computer Science, 2020, 47(11A): 244-247.
[14] LI Zong-min, LI Si-yuan, LIU Yu-jie, LI Hua. Sketch-based Image Retrieval Based on Attention Model [J]. Computer Science, 2020, 47(11): 199-204.
[15] SUN Wei, ZHAO Yu-pu. Enhanced Rotation Invariant LBP Algorithm and Its Application in Image Retrieval [J]. Computer Science, 2019, 46(7): 263-267.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!