Computer Science ›› 2020, Vol. 47 ›› Issue (5): 190-197.doi: 10.11896/jsjkx.190700128

• Artificial Intelligence • Previous Articles     Next Articles

Analyzing Latent Representation of Deep Neural Networks Based on Feature Visualization

SHANG Jun-yuan, YANG Le-han, HE Kun   

  1. School of Computer Science and Technology,Huazhong University of Science and Technology,Wuhan 430074,China
  • Received:2019-07-17 Online:2020-05-15 Published:2020-05-19
  • About author:SHANG Jun-yuan,born in 1994,M.S.candidate.His main research interests include machine learning and deep learning.
    HE Kun,born in 1972,professor and Ph.D.supervisor.Her main research interest include machine learning,deep learning,and optimization algorithms.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China(61772219) and Fundamental Research Funds for the Central Universities of Ministry of Education of China (2019kfyXKJC021).

Abstract: The working mechanism of deep neural networks can be intuitively uncovered by visualization technique.Visualizing deep neural networks can provide the interpretability on the decision made by the black box model,which is critically important in many fields,such as medical diagnosis and autopilot.Current existing works are mostly based on the activation maximization technique,which optimizes the input,the hidden feature map or the original image,in condition to the neuron that we want to observe.Qualitatively,the change in the input value can be taken as explanation when the neuron has reached nearly the maximum activation value.However,such method lacks the quantitative analysis of deep neural networks.To fill this gap,this paper proposes two meta methods,namely,structure visualization and rule-based visualization.Structure visualization works by visualizing from the shallow layers to the deep layers,and find that neurons in shallow layers learn global characteristics while neurons in deep layers learn more specific features.The rule-based visualization includes intersection and difference selection rule,and it is helpful to find the existence of shared neurons and inhibition neurons that learns the common features of different categories and inhibits unrelated features respectively.Experiments on two representative deep networks,namely the convolutional network VGG and the residual network ResNet,by using ImageNet and COCO datasets.Quantitative analysis shows that ResNet and VGG are highly sparse in representation.Thus,by removing some low activation-value “noisy” neurons,the networks can keep or even improve the classification accuracy.This paper discovers the Latent representation of deep neural networks by visualizing and quantitatively analyzing hidden features,thus providing guidance and reference for the design of high-performance deep neural networks.

Key words: Deep neural network, Feature visualization, Inhibition neuron, Internal representation, Shared neuron

CLC Number: 

  • TP83
[1]MCCULLOCH D E,PITTS W.A logical calculus of ideas immanent in nervous activity [J].Bulletin of Mathematical Biophysics,1943,5:115-133.
[2]RUMELHART D E,HINTON G E,WILLIAMS R J.Learning representations by back-propagating errors [J].Nature,1986,323:533-536.
[3]HINTON G E,SALAKHUTDINOV G E.Reducing the dimensionality of data with neural networks [J].Science,2006,313(5786):504-507.
[4]JIAO L C,YANG S Y,LIU F,et al.Seventy Years beyond Neural Networks:Retrospect and Prospect[J].Chinese Journal of Computers,2016,39(8):1697-1716.
[5]ZHOU F Y,JIN L P,DONG J.Review of Convolutional Neural Network[J].Chinese Journal of Computers,2017,40(6):1229-1251.
[6]CIREŞAN D C,MEIER U,GAMBARDELLA L M,et al.Deep,big,simple neural nets for handwritten digit recognition [J].Neural Computation,2010,22(12):3207-3220.
[7]FARABET C,COUPRIE C,NAJMAN L,et al.Learning hierarchical features for scene labeling [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2013,35(8):1915-1929.
[8]ZHAO R,OUYANG W,LI H S,et al.Saliency detection bymulti-context deep learning[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2015:1265-1274.
[9]MOHAMED A,DAHL G E,HINTON G E.Acoustic modeling using deep belief networks [J].IEEE Transactions on Audio,Speech,and Language Processing,2012,20(1):14-22.
[10]RUSSAKOVSKY O,DENG J,SU H,et al.ImageNet large scale visual recognition challenge [J].International Journal of Computer Vision,2015,115(3):211-252.
[11]KRIZHEVSKY A,SUTSKEVER I,HINTON G E.ImageNet classification with deep convolutional neural networks[C]//Advances in neural information processing systems.2012:1097-1105.
[12]SIMONYAN K,ZISSERMAN A.Very deep convolutional networks for large-scale image recognition [C]//3rd International Conference on Learning Representations (ICLR).2015.
[13]HE K M,ZHANG X,REN S Q,et al.Deep residual learning for image recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770-778.
[14]SRIVASTAVA N,HINTON G E,KRIZHEVSKY A,et al.Dropout:a simple way to prevent neural networks from overfitting [J].Journal of Machine Learning Research,2014,15(1):1929-1958.
[15]IOFFE S,SZEGEDY C.Batch normalization:Accelerating deep network training by reducing internal covariate shift [C]//Proceedings of the 32nd International Conference on Machine Learning (ICML).2015.
[16]YOSINSKI J,CLUNE J,NGUYEN A M,et al.Understanding neural networks through deep visualization [C]//Deep Learning Workshop,International Conference on Machine Learning (ICML).Lille,France,2015.
[17]ZEILER M D,FERGUS R.Visualizing and understanding convolutional networks[C]//European Conference on Computer Vision (ECCV).Springer International Publishing,2014:818-833.
[18]ERHAN D,BENGIO Y,COURVILLE A,et al.Visualizinghigher-layer features of a deep network:Technical Report 1341[R].University of Montreal,2009.
[19]ZEILER M D,TAYLOR G W,FERGUS R.Adaptive deconvolutional networks for mid and high level feature learning[C]//International Conference on Computer Vision (ICCV).2011.
[20]SIMONYAN K,VEDALDI A,ZISSERMAN A.Deep insideconvolutional networks:Visualising image classification models and saliency maps [C]//2nd International Conference on Learning Representations (ICLR).2014.
[21]ZINTGRAF L M,COHEN T S,ADEL T,et al.Visualizing deep neural network decisions:Prediction difference analysis [C]//5th International Conference on Learning Representations (ICLR).2017.
[22]ZHOU B L,KHOSLA A,LAPEDRIZA A,et al.Learning deep features for discriminative localization[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2016:2921-2929.
[23]MAATEN L V D,HINTON G E.Visualizing data using t-SNE [J].Journal of Machine Learning Research,2008(9):2579-2605.
[24]DOSOVITSKIY A,BROX T.Inverting visual representationswith convolutional networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2016:4829-4837.
[25]FONG R,VEDALDI A.Net2Vec:Quantifying and Explaining How Concepts Are Encoded by Filters in Deep Neural Networks[C]//2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2018.
[26]FONG R C,VEDALDI A.Interpretable Explanations of Black Boxes by Meaningful Perturbation[C]//Proceedings of the International Conference on Computer Vision (ICCV).2017:3449-3457.
[27]NGUYEN A,YOSINSKI J,CLUNE J.Understanding NeuralNetworks via Feature Visualization:A Survey[C]//Explainable AI:Interpreting,Explaining and Visualizing Deep Learning.Cham:Springer,2019:55-76.
[28]CHEN J B,SONG L,WAINWRIGHT M J,et al.Learning to explain:An information-theoretic perspective on model interpretation[C]//Proceedings of the 35nd International Conference on Machine Learning (ICML).2018.
[29]SAMEK W,BINDER A,MONTAVON G,et al.Evaluating the visualization of what a deep neural network has learned [J].IEEE Transactions on Neural Networks and Learning Systems,2016,28(11):2660-2673.
[30]LIN T Y,MAIRE M,BELONGIE S J,et al.Microsoft COCO:Common objects in context[C]//European Conference on Computer Vision (ECCV).Cham:Springer,2014:740-755.
[31]ABADI M,AGARWAL A,BARHAM P,et al.Tensorflow:Large-scale machine learning on heterogeneous distributed systems [J].arXiv:1603.04467,2016.
[32]LECUN Y,BOTTOU L,BENGIO Y,et al.Gradient-basedlearning applied to document recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324.
[1] WEI Hui, CHEN Ze-mao, ZHANG Li-qiang. Anomaly Detection Framework of System Call Trace Based on Sequence and Frequency Patterns [J]. Computer Science, 2022, 49(6): 350-355.
[2] JIAO Xiang, WEI Xiang-lin, XUE Yu, WANG Chao, DUAN Qiang. Automatic Modulation Recognition Based on Deep Learning [J]. Computer Science, 2022, 49(5): 266-278.
[3] GAO Jie, LIU Sha, HUANG Ze-qiang, ZHENG Tian-yu, LIU Xin, QI Feng-bin. Deep Neural Network Operator Acceleration Library Optimization Based on Domestic Many-core Processor [J]. Computer Science, 2022, 49(5): 355-362.
[4] FAN Hong-jie, LI Xue-dong, YE Song-tao. Aided Disease Diagnosis Method for EMR Semantic Analysis [J]. Computer Science, 2022, 49(1): 153-158.
[5] ZHOU Xin, LIU Shuo-di, PAN Wei, CHEN Yuan-yuan. Vehicle Color Recognition in Natural Traffic Scene [J]. Computer Science, 2021, 48(6A): 15-20.
[6] LIU Dong, WANG Ye-fei, LIN Jian-ping, MA Hai-chuan, YANG Run-yu. Advances in End-to-End Optimized Image Compression Technologies [J]. Computer Science, 2021, 48(3): 1-8.
[7] PAN Yu, ZOU Jun-hua, WANG Shuai-hui, HU Gu-yu, PAN Zhi-song. Deep Community Detection Algorithm Based on Network Representation Learning [J]. Computer Science, 2021, 48(11A): 198-203.
[8] MA Lin, WANG Yun-xiao, ZHAO Li-na, HAN Xing-wang, NI Jin-chao, ZHANG Jie. Network Intrusion Detection System Based on Multi-model Ensemble [J]. Computer Science, 2021, 48(11A): 592-596.
[9] LIU Tian-xing, LI Wei, XU Zheng, ZHANG Li-hua, QI Xiao-ya, GAN Zhong-xue. Monte Carlo Tree Search for High-dimensional Continuous Control Space [J]. Computer Science, 2021, 48(10): 30-36.
[10] ZHANG Yan-mei, LOU Yin-cheng. Deep Neural Network Based Ponzi Scheme Contract Detection Method [J]. Computer Science, 2021, 48(1): 273-279.
[11] LV Ze-yu, LI Ji-xuan, CHEN Ru-Jian and CHEN Dong-ming. Research on Prediction of Re-shopping Behavior of E-commerce Customers [J]. Computer Science, 2020, 47(6A): 424-428.
[12] DING Zi-ang, LE Cao-wei, WU Ling-ling and FU Ming-lei. PM2.5 Concentration Prediction Method Based on CEEMD-Pearson and Deep LSTM Hybrid Model [J]. Computer Science, 2020, 47(6A): 444-449.
[13] TANG Guo-qiang,GAO Da-qi,RUAN Tong,YE Qi,WANG Qi. Clinical Electronic Medical Record Named Entity Recognition Incorporating Language Model and Attention Mechanism [J]. Computer Science, 2020, 47(3): 211-216.
[14] FAN Wei, LIU Ting, HUANG Rui, GUO Qing, ZHANG Bao. Low-level CNN Feature Aided Image Instance Segmentation [J]. Computer Science, 2020, 47(11): 186-191.
[15] KONG Fan-yu, ZHOU Yu-feng, CHEN Gang. Traffic Flow Prediction Method Based on Spatio-Temporal Feature Mining [J]. Computer Science, 2019, 46(7): 322-326.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!