Computer Science ›› 2022, Vol. 49 ›› Issue (3): 134-143.doi: 10.11896/jsjkx.210100001

• Database & Big Data & Data Science • Previous Articles     Next Articles

Self-supervised Deep Clustering Algorithm Based on Self-attention

HAN Jie1, CHEN Jun-fen1, LI Yan2, ZHAN Ze-cong1   

  1. 1 Hebei Key Laboratory of Machine Learning and Computational Intelligence,College of Mathematics and Information Sciences,Hebei University,Baoding,Hebei 071002,China
    2 School of Applied Mathematics,Beijing Normal University Zhuhai,Zhuhai,Guangdong 519087,China
  • Received:2021-01-03 Revised:2021-07-08 Online:2022-03-15 Published:2022-03-15
  • About author:HAN Jie,born in 1996,postgraduate.Her main research interests include image clustering and machine learning.
    CHEN Jun-fen,born in 1976,Ph.D,associate professor,master supervisor,is a member of China Computer Federation.Her main research interests include data mining,machine learning and image processing.
  • Supported by:
    Hebei Province Introduction of Studying Abroad Talent Funded Project(C20200302),Natural Science Foundation of Hebei Province(F2018201096),Natural Science Foundation of Guangdong Province(2018A0303130026) and Social Science Foundation of Hebei Province(HB20TQ005).

Abstract: In recent years,deep clustering methods using joint optimization strategy,such as DEC (deep embedding clustering) and DDC (deep denoising clustering) algorithms,have made great progress in image clustering that heavily related to features representation ability of deep networks,and brought certain degree breakthroughs in clustering performances.The quality of feature extraction directlyaffects the subsequent clustering tasks.However,the generalization abilities of these methods are not satisfied,exactly as different network structures are used in different datasets to guarantee the clustering performance.In addition,there is a quite larger space to enhance clustering performances compared to classification performances.To this end,a self-supervised deep clustering (SADC) method based on self-attention is proposed.Firstly,a deep convolutional autoencoder is designed to extract features,and noisy images are employed to enhance the robustness of the network.Secondly,self-attention mechanism is combined with the proposed network to capture useful features for clustering.At last,the trained encoder combines with K-means algorithm to form a deep clustering model for feature representation and clustering assignment,and iteratively updates parameters to improve the clustering accuracy and generalization ability of the proposed network.The proposed clustering method is verified on 6 traditional image datasets and compared with the deep clustering algorithms DEC and DDC.Experimental results show that the proposed SADC can provide better clustering results,and is comparable to the state-of-the-art clustering algorithms.Overall,the unified network structure ensures the clustering accuracy and simultaneously reducing computational complexity of the deep clustering algorithms.

Key words: Computational complexity, Deep convolutional autoencoder, Features representation, Image clustering, Self-attention

CLC Number: 

  • TP181
[1]ASANO Y M,RUPPRECHT C,VEDALDI A.Self-labelling via simultaneous clustering and repre-sentation learning[C]//Proceedings of the International Conference on Learning Representations (ICLR).2020:1-22.
[2]WANG C,BAI X,DU J.Diffuse Interface Based Unsupervised Images Clustering Algorithm[J].Computer Science,2020,47(5):149-153.
[3]MCCULLOCH W S,PITTS W H.A logical calculus of the ideas immanent in nervous activity[J].The Bulletin of Mathematical Biophysics,1988,5:115-133.
[4]LECUN Y,BOTTOU L.Gradient-based learning applied to do-cument recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324.
[5]ZHAO H,JIA J,KOLTN V.Exploring self-attention for image recognition[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2020:10076-10085.
[6]SHANG L,LU Z,LI H.Neural Responding Machine for Short-Text Conversation[C]//Proceedings of the Meeting of the Association for Computational Linguistics.2015.
[7]VASWANI A,SHAZEER N,PARMAR N,et al.Attention Is All You Need[C]//Proceedings of the Advances in Neural Information Processing Systems(NeuIPS).2017:5998-6008.
[8]XU K,BA J,KIROS R,et al.Show,Attend and Tell:NeuralImage Caption Generation with Visual Attention[C]//Procee-dings of the International Conference on Machine Learning(ICML).2015:2048-2057.
[9]HUANG P,HUANG Y,WANG W,et al.Deep embedding network for clustering[C]//Proceedings of the IEEE International Conference on Pattern Recognition(ICPR).2014:1532-1537.
[10]YANG B,FU X.Towards k-means-friendly spaces:Simul-taneous deep earning and clustering[C]//Proceedings of the International Conference on Machine Learning.2017:3861-3870.
[11]XIE J,GIRSHICK R,FARHADI A.Unsupervised Deep Em-bedding for Clustering Analysis[C]//Proceedings of the International Conference on Machine Learning(ICML).2016:478-487.
[12]LI F,QIAO H,ZHANG B.Discriminatively Boosted ImageClustering with Fully Convolutional Auto-Encoders[J].Pattern Recognition,2018,83:161-173.
[13]CHEN J,ZHANG M,ZHAO J.A Deep Clustering AlgorithmBased on Denoising and Self-attention[J].Computer Science and Techno-logy,2020,15(9):1117-1727.
[14]XIE J,HOU Q,CAO J.Image Clustering Algorithms by Deep Convolutional Autoencoders[J].Computer Science and Techno-logy,2019,13(4):586-595.
[15]LI B, PI D, CUI L,et al.DNC:A Deep Neural Network-based Clustering-oriented Network Embedding Algorithm[J].Journal of Network and Computer Applications,2020,173:102854.
[16]DING Y,WEI H,PAN Z,et al.Survey of Net-work Representation Learning[J].Computer Science,2020,47(9):52-59.
[17]HE K,SUN J.Convolutional neural networks at constrainedtime cost[C]//Proceedings of the IEEE Conference on Compu-ter Vision and Pattern Recognition.2015:5353-5360.
[18]HU W,MIYATO T,TOKUI S,et al.Learning Discrete Representations via Information Maximizing Self-Augmented Training[C]//Proceedings of the 34th International Conference on Machine Learning (ICML).2017:1-15.
[19]BALDI P.Autoencoders,Unsupervised learning and deep architectures[C]//Proceedings of the International Conference on Unsupervised and Transfer Learning Workshop.2011:37-50.
[20]VICENT P,LAROCHELLE H,LAJOIE I,et al.Stacked De-noising Autoencoders:Learning Useful Representations in a Deep Network with a Local Denoising Criterion[J].Journal of Machine Learning Research,2010,11(12):3371-3408.
[21]DIZAJI K G,HERANDI A,HUANG H. Deep clustering viajoint convolutional autoencoder embedding and relative entropy minimization[C]//Proceedings of the IEEE International Conference on Computer Vision.2017:5736-5745.
[22]LIN Z,FENG M,SANTOS C N D,et al.A Structured Self-attentive Sentence Embedding[C]//Proceedings of the International Conference on Learning Representations.2017:1-15.
[23]WANG X,GIRSHICK R,GUPTA A,et al.Non-local Neuralnetworks[C]//Proceedings of the IEEE Conference on Compu-ter Vision and Pattern Recognition.2018:7794-7803.
[24]ZHANG H,GOODFELLOW I,METAXAS D,et al.Self-Attention Generative Adversarial Networks[C]//International Conference on Machine Learning.PMLR,2019:7354-7363.
[25]BOTTOU L,CURTIS F E,NOCEDAL J.Optimization methods for large-scale machine learning[J].Siam Review,2018,60(2):223-311.
[26]RUDER S.An overview of gradient descent optimization algo-rithms[J].arXiv:1609.04747,2016.
[27]RUMELHARTT D,HINTON G,WILLIAMS R.Learning rep-resentations by back-propagating errors[J].Nature,1986,323(6088):533-536.
[28]LECUN Y,BOTTOU L.Gradient-based learning applied to do-cument recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324.
[29]NAIR V,HINTON G.Rectified linear units i-mprove restricted boltzmann machines[C]//Proceedings of the 27th International Conference on Machine Learning.Madison:Omni press,2010:807-814.
[30]KINGMA D,BA L.Adam:A method for sto-chastic optimiza-tion[C]//Proceedings of the 3rd International Conference on Learning Representations.2015:1-15.
[31]LAURENS V,HINTON G.Visualizing Data using t-SNE[J].Journal of Machine Learning Research,2008,9(2605):2579-2605.
[1] WU Zi-yi, LI Shao-mei, JIANG Meng-han, ZHANG Jian-peng. Ontology Alignment Method Based on Self-attention [J]. Computer Science, 2022, 49(9): 215-220.
[2] FANG Yi-qiu, ZHANG Zhen-kun, GE Jun-wei. Cross-domain Recommendation Algorithm Based on Self-attention Mechanism and Transfer Learning [J]. Computer Science, 2022, 49(8): 70-77.
[3] CHEN Kun-feng, PAN Zhi-song, WANG Jia-bao, SHI Lei, ZHANG Jin. Moderate Clothes-Changing Person Re-identification Based on Bionics of Binocular Summation [J]. Computer Science, 2022, 49(8): 165-171.
[4] JIN Fang-yan, WANG Xiu-li. Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM [J]. Computer Science, 2022, 49(7): 179-186.
[5] ZHANG Jia-hao, LIU Feng, QI Jia-yin. Lightweight Micro-expression Recognition Architecture Based on Bottleneck Transformer [J]. Computer Science, 2022, 49(6A): 370-377.
[6] HUANG Hua-wei, LI Chun-hua. Security Analysis of A Key Exchange Protocol Based on Tropical Semi-ring [J]. Computer Science, 2022, 49(6A): 571-574.
[7] ZHAO Dan-dan, HUANG De-gen, MENG Jia-na, DONG Yu, ZHANG Pan. Chinese Entity Relations Classification Based on BERT-GRU-ATT [J]. Computer Science, 2022, 49(6): 319-325.
[8] HU Yan-li, TONG Tan-qian, ZHANG Xiao-yu, PENG Juan. Self-attention-based BGRU and CNN for Sentiment Analysis [J]. Computer Science, 2022, 49(1): 252-258.
[9] YOU Ling, GUAN Zhang-jun. Low-complexity Subcarrier Allocation Algorithm for Underwater OFDM Acoustic CommunicationSystems [J]. Computer Science, 2021, 48(6A): 387-391.
[10] CHEN Yang, WANG Jin-liang, XIA Wei, YANG Hao, ZHU Run, XI Xue-feng. Footprint Image Clustering Method Based on Automatic Feature Extraction [J]. Computer Science, 2021, 48(6A): 255-259.
[11] HU De-feng, ZHANG Chen-xi, WANG Shi-tao, ZHAO Qin-pei, LI Jiang-feng. Intelligent Prediction Model of Tool Wear Based on Deep Signal Processing and Stacked-ResGRU [J]. Computer Science, 2021, 48(6): 175-183.
[12] WANG Xi, ZHANG Kai, LI Jun-hui, KONG Fang, ZHANG Yi-tian. Generation of Image Caption of Joint Self-attention and Recurrent Neural Network [J]. Computer Science, 2021, 48(4): 157-163.
[13] ZHOU Xiao-shi, ZHANG Zi-wei, WEN Juan. Natural Language Steganography Based on Neural Machine Translation [J]. Computer Science, 2021, 48(11A): 557-564.
[14] ZHANG Shi-hao, DU Sheng-dong, JIA Zhen, LI Tian-rui. Medical Entity Relation Extraction Based on Deep Neural Network and Self-attention Mechanism [J]. Computer Science, 2021, 48(10): 77-84.
[15] YU Wen-jia, DING Shi-fei. Conditional Generative Adversarial Network Based on Self-attention Mechanism [J]. Computer Science, 2021, 48(1): 241-246.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!