Computer Science ›› 2024, Vol. 51 ›› Issue (6A): 230400112-7.doi: 10.11896/jsjkx.230400112

• Artificial Intelligenc • Previous Articles     Next Articles

DRSTN:Deep Residual Soft Thresholding Network

CAO Yan, ZHU Zhenfeng   

  1. School of Computer and Artificial Intelligence,Zhengzhou University,Zhengzhou 450001,China
  • Published:2024-06-06
  • About author:CAO Yan,born in 1993,postgraduate.His main research interests include machine learning and transfer learning.
    ZHU Zhenfeng,born in 1980,Ph.D,associate professor,is a member of CCF(No.18743M).His main research interests include data mining,machine learning and computer vision.
  • Supported by:
    National Natural Science Foundation of China(62176239).

Abstract: When using neural network models such as deep residuals to classify images,some important features lost during feature extraction will affect the classification performance of the model.The black box problem brought about by the “end-to-end” learning mode of neural network can also limit its application and development in many fields.In addition,neural network models often require longer training time than traditional methods.In order to improve the classification effect and training efficiency of the deep residual networks,this paper introduces the model transfer method and soft thresholding method,proposes the deep residual soft thresholding network(DRSTN) network,and fine-tunes the network structure to generate different versions DRSTN network.The performance of the DRSTN networks benefit from the organic integration of three aspects:1)Visualize the feature extraction of the network through the gradients-weighted class activation mapping(Grad-CAM) method,and select further optimized ones based on the visualization results.2)Based on model transfer,researchers do not need to build a model from scratch,and can directly optimize the existing models,which can save a lot of training time.3) Soft thresholding,as a nonlinear transformation layer,is embedded into the deep residual network architecture to eliminate irrelevant features in samples.Experimental results show that under the same training conditions,the classification accuracy of the DRSTN_KS(3*3)_RB(2:2:2) network on the CIFAR-10 dataset is 15.5%,8.8% and 10.9% higher than that of SKNet-18,ResNet18 and ConvNeXt_tiny networks,respectively.The network also has a certain degree of generalization.It can achieve rapid transfer on MNIST and Fashion MNIST datasets,and the classification accuracy reaches 99.06% and 93.15% respectively.

Key words: Transfer learning, Residual network, Gradient weighted class activation mapping, Soft thresholding method, Image classification

CLC Number: 

  • TP391
[1]LECUN Y,BENGIO Y,HINTON G.Deep learning[J].Nature,2015,521(7553):436-444.
[2]HASSANZADEH T,ESSAM D,SARKER R.EvoDCNN:Anevolutionary deep convolutional neural network for image classification[J].Neurocomputing,2022,488:271-283.
[3]FADLULLAH Z M,TANG F,MAO B,et al.State-of-the-art deep learning:Evolving machine intelligence toward tomorrow’s intelligent network traffic control systems[J].IEEE Communications Surveys & Tutorials,2017,19(4):2432-2455.
[4]ZHANG Z,GEIGER J,POHJALAINEN J,et al.Deep learning for environmentally robust speech recognition:An overview of recent developments[J].ACM Transactions on Intelligent Systems and Technology,2018,9(5):1-28.
[5]DENG J,GUO J,XUE N,et al.Arcface:Additive angular margin loss for deep face recognition[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:4690-4699.
[6]DU J,CHEN Y H,ZHANG L,et al.Low-power expression recognition based on improved deep residual network[J].Computer Science,2018,45(9):303-307.
[7]KRIZHEVSKY A,SUTSKEVER I,HINTON G E.Imagenetclassification with deep convolutional neural networks[J].Communications of the ACM,2017,60(6):84-90.
[8]SIMONYAN K,ZISSERMAN A.Very deep convolutional networks for large-scale image recognition[J].arXiv:1409.1556,2015.
[9]SZEGEDY C,LIU W,JIA Y,et al.Going deeper with convolutions[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2015:1-9.
[10]HE K,ZHANG X,REN S,et al.Deep residual learning for image recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770-778.
[11]HUANG G,LIU Z,VAN DER MAATEN L,et al.Densely con-nected convolutional networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:4700-4708.
[12]HASSANPOUR M,MALEK H.Document image classificationusing SqueezeNet convolutional neural network[C]//2019 5th Iranian Conference on Signal Processing and Intelligent Systems.IEEE,2019:1-4.
[13]LI X,WANG W,HU X,et al.Selective kernel networks[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2019:510-519.
[14]DOSOVITSKIY A,BEYER L,KOLESNIKOV A,et al.Animage is worth 16x16 words:Transformers for image recognition at scale[J].arXiv:2010.11929,2021.
[15]LIU Z,LIN Y,CAO Y,et al.Swin transformer:Hierarchical vi-sion transformer using shifted windows[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision.2021:10012-10022.
[16]LIU Z,MAO H,WU C Y,et al.A convnet for the 2020s[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2022:11976-11986.
[17]SELVARAJU R R,COGSWELL M,DAS A,et al.Grad-cam:Visual explanations from deep networks via gradient-based localization[C]//Proceedings of the IEEE International Confe-rence on Computer Vision.2017:618-626.
[18]PAN S J,YANG Q.A survey on transfer learning[J].IEEE Transactions on Knowledge and Data Engineering,2010,22(10):1345-1359.
[19]ISOGAWA K,IDA T,SHIODERA T,et al.Deep shrinkageconvolutional neural network for adaptive noise reduction[J].IEEE Signal Processing Letters,2017,25(2):224-228.
[1] ZHANG Le, YU Ying, GE Hao. Mural Inpainting Based on Fast Fourier Convolution and Feature Pruning Coordinate Attention [J]. Computer Science, 2024, 51(6A): 230400083-9.
[2] LI Xinrui, ZHANG Yanfang, KANG Xiaodong, LI Bo, HAN Junling. Intelligent Diagnosis of Brain Tumor with MRI Based on Ensemble Learning [J]. Computer Science, 2024, 51(6A): 230600043-7.
[3] ZHANG Xinrui, YANG Jian, WANG Zhan. Thai Speech Synthesis Based on Cross-language Transfer Learning and Joint Training [J]. Computer Science, 2024, 51(6A): 230500174-7.
[4] SU Ruqi, BIAN Xiong, ZHU Songhao. Few-shot Images Classification Based on Clustering Optimization Learning [J]. Computer Science, 2024, 51(6A): 230300227-7.
[5] LYU Yiming, WANG Jiyang. Iron Ore Image Classification Method Based on Improved Efficientnetv2 [J]. Computer Science, 2024, 51(6A): 230600212-6.
[6] HOU Lei, LIU Jinhuan, YU Xu, DU Junwei. Review of Graph Neural Networks [J]. Computer Science, 2024, 51(6): 282-298.
[7] WANG Jiahao, FU Yifu, FENG Hainan, REN Yuheng. Indoor Location Algorithm in Dynamic Environment Based on Transfer Learning [J]. Computer Science, 2024, 51(5): 277-283.
[8] ZENG Ruiren, XIE Jiangtao, LI Peihua. Global Covariance Pooling Based on Fast Maximum Singular Value Power Normalization [J]. Computer Science, 2024, 51(4): 254-261.
[9] LUO Zeyang, TIAN Hua, DOU Yingtong, LI Manwen, ZHANG Zehua. Fake Review Detection Based on Residual Networks Fusion of Multi-relationship Review Features [J]. Computer Science, 2024, 51(4): 314-323.
[10] WANG Wenjie, YANG Yan, JING Lili, WANG Jie, LIU Yan. LNG-Transformer:An Image Classification Network Based on Multi-scale Information Interaction [J]. Computer Science, 2024, 51(2): 189-195.
[11] ZHANG Feng, HUANG Shixin, HUA Qiang, DONG Chunru. Novel Image Classification Model Based on Depth-wise Convolution Neural Network andVisual Transformer [J]. Computer Science, 2024, 51(2): 196-204.
[12] YANG Lin, YANG Jian, CAI Haoran, LIU Cong. Vietnamese Speech Synthesis Based on Transfer Learning [J]. Computer Science, 2023, 50(8): 118-124.
[13] LI Fan, JIA Dongli, YAO Yumin, TU Jun. Graph Neural Network Few Shot Image Classification Network Based on Residual and Self-attention Mechanism [J]. Computer Science, 2023, 50(6A): 220500104-5.
[14] WANG Tianran, WANG Qi, WANG Qingshan. Transfer Learning Based Cross-object Sign Language Gesture Recognition Method [J]. Computer Science, 2023, 50(6A): 220300232-5.
[15] HU Mingyang, GUO Yan, JIN Yangshuang. PSwin:Edge Detection Algorithm Based on Swin Transformer [J]. Computer Science, 2023, 50(6): 194-199.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!