Computer Science ›› 2020, Vol. 47 ›› Issue (5): 190-197.doi: 10.11896/jsjkx.190700128
• Artificial Intelligence • Previous Articles Next Articles
SHANG Jun-yuan, YANG Le-han, HE Kun
CLC Number:
[1]MCCULLOCH D E,PITTS W.A logical calculus of ideas immanent in nervous activity [J].Bulletin of Mathematical Biophysics,1943,5:115-133. [2]RUMELHART D E,HINTON G E,WILLIAMS R J.Learning representations by back-propagating errors [J].Nature,1986,323:533-536. [3]HINTON G E,SALAKHUTDINOV G E.Reducing the dimensionality of data with neural networks [J].Science,2006,313(5786):504-507. [4]JIAO L C,YANG S Y,LIU F,et al.Seventy Years beyond Neural Networks:Retrospect and Prospect[J].Chinese Journal of Computers,2016,39(8):1697-1716. [5]ZHOU F Y,JIN L P,DONG J.Review of Convolutional Neural Network[J].Chinese Journal of Computers,2017,40(6):1229-1251. [6]CIREŞAN D C,MEIER U,GAMBARDELLA L M,et al.Deep,big,simple neural nets for handwritten digit recognition [J].Neural Computation,2010,22(12):3207-3220. [7]FARABET C,COUPRIE C,NAJMAN L,et al.Learning hierarchical features for scene labeling [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2013,35(8):1915-1929. [8]ZHAO R,OUYANG W,LI H S,et al.Saliency detection bymulti-context deep learning[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2015:1265-1274. [9]MOHAMED A,DAHL G E,HINTON G E.Acoustic modeling using deep belief networks [J].IEEE Transactions on Audio,Speech,and Language Processing,2012,20(1):14-22. [10]RUSSAKOVSKY O,DENG J,SU H,et al.ImageNet large scale visual recognition challenge [J].International Journal of Computer Vision,2015,115(3):211-252. [11]KRIZHEVSKY A,SUTSKEVER I,HINTON G E.ImageNet classification with deep convolutional neural networks[C]//Advances in neural information processing systems.2012:1097-1105. [12]SIMONYAN K,ZISSERMAN A.Very deep convolutional networks for large-scale image recognition [C]//3rd International Conference on Learning Representations (ICLR).2015. [13]HE K M,ZHANG X,REN S Q,et al.Deep residual learning for image recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770-778. [14]SRIVASTAVA N,HINTON G E,KRIZHEVSKY A,et al.Dropout:a simple way to prevent neural networks from overfitting [J].Journal of Machine Learning Research,2014,15(1):1929-1958. [15]IOFFE S,SZEGEDY C.Batch normalization:Accelerating deep network training by reducing internal covariate shift [C]//Proceedings of the 32nd International Conference on Machine Learning (ICML).2015. [16]YOSINSKI J,CLUNE J,NGUYEN A M,et al.Understanding neural networks through deep visualization [C]//Deep Learning Workshop,International Conference on Machine Learning (ICML).Lille,France,2015. [17]ZEILER M D,FERGUS R.Visualizing and understanding convolutional networks[C]//European Conference on Computer Vision (ECCV).Springer International Publishing,2014:818-833. [18]ERHAN D,BENGIO Y,COURVILLE A,et al.Visualizinghigher-layer features of a deep network:Technical Report 1341[R].University of Montreal,2009. [19]ZEILER M D,TAYLOR G W,FERGUS R.Adaptive deconvolutional networks for mid and high level feature learning[C]//International Conference on Computer Vision (ICCV).2011. [20]SIMONYAN K,VEDALDI A,ZISSERMAN A.Deep insideconvolutional networks:Visualising image classification models and saliency maps [C]//2nd International Conference on Learning Representations (ICLR).2014. [21]ZINTGRAF L M,COHEN T S,ADEL T,et al.Visualizing deep neural network decisions:Prediction difference analysis [C]//5th International Conference on Learning Representations (ICLR).2017. [22]ZHOU B L,KHOSLA A,LAPEDRIZA A,et al.Learning deep features for discriminative localization[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2016:2921-2929. [23]MAATEN L V D,HINTON G E.Visualizing data using t-SNE [J].Journal of Machine Learning Research,2008(9):2579-2605. [24]DOSOVITSKIY A,BROX T.Inverting visual representationswith convolutional networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2016:4829-4837. [25]FONG R,VEDALDI A.Net2Vec:Quantifying and Explaining How Concepts Are Encoded by Filters in Deep Neural Networks[C]//2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).2018. [26]FONG R C,VEDALDI A.Interpretable Explanations of Black Boxes by Meaningful Perturbation[C]//Proceedings of the International Conference on Computer Vision (ICCV).2017:3449-3457. [27]NGUYEN A,YOSINSKI J,CLUNE J.Understanding NeuralNetworks via Feature Visualization:A Survey[C]//Explainable AI:Interpreting,Explaining and Visualizing Deep Learning.Cham:Springer,2019:55-76. [28]CHEN J B,SONG L,WAINWRIGHT M J,et al.Learning to explain:An information-theoretic perspective on model interpretation[C]//Proceedings of the 35nd International Conference on Machine Learning (ICML).2018. [29]SAMEK W,BINDER A,MONTAVON G,et al.Evaluating the visualization of what a deep neural network has learned [J].IEEE Transactions on Neural Networks and Learning Systems,2016,28(11):2660-2673. [30]LIN T Y,MAIRE M,BELONGIE S J,et al.Microsoft COCO:Common objects in context[C]//European Conference on Computer Vision (ECCV).Cham:Springer,2014:740-755. [31]ABADI M,AGARWAL A,BARHAM P,et al.Tensorflow:Large-scale machine learning on heterogeneous distributed systems [J].arXiv:1603.04467,2016. [32]LECUN Y,BOTTOU L,BENGIO Y,et al.Gradient-basedlearning applied to document recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324. |
[1] | WEI Hui, CHEN Ze-mao, ZHANG Li-qiang. Anomaly Detection Framework of System Call Trace Based on Sequence and Frequency Patterns [J]. Computer Science, 2022, 49(6): 350-355. |
[2] | JIAO Xiang, WEI Xiang-lin, XUE Yu, WANG Chao, DUAN Qiang. Automatic Modulation Recognition Based on Deep Learning [J]. Computer Science, 2022, 49(5): 266-278. |
[3] | GAO Jie, LIU Sha, HUANG Ze-qiang, ZHENG Tian-yu, LIU Xin, QI Feng-bin. Deep Neural Network Operator Acceleration Library Optimization Based on Domestic Many-core Processor [J]. Computer Science, 2022, 49(5): 355-362. |
[4] | FAN Hong-jie, LI Xue-dong, YE Song-tao. Aided Disease Diagnosis Method for EMR Semantic Analysis [J]. Computer Science, 2022, 49(1): 153-158. |
[5] | ZHOU Xin, LIU Shuo-di, PAN Wei, CHEN Yuan-yuan. Vehicle Color Recognition in Natural Traffic Scene [J]. Computer Science, 2021, 48(6A): 15-20. |
[6] | LIU Dong, WANG Ye-fei, LIN Jian-ping, MA Hai-chuan, YANG Run-yu. Advances in End-to-End Optimized Image Compression Technologies [J]. Computer Science, 2021, 48(3): 1-8. |
[7] | PAN Yu, ZOU Jun-hua, WANG Shuai-hui, HU Gu-yu, PAN Zhi-song. Deep Community Detection Algorithm Based on Network Representation Learning [J]. Computer Science, 2021, 48(11A): 198-203. |
[8] | MA Lin, WANG Yun-xiao, ZHAO Li-na, HAN Xing-wang, NI Jin-chao, ZHANG Jie. Network Intrusion Detection System Based on Multi-model Ensemble [J]. Computer Science, 2021, 48(11A): 592-596. |
[9] | LIU Tian-xing, LI Wei, XU Zheng, ZHANG Li-hua, QI Xiao-ya, GAN Zhong-xue. Monte Carlo Tree Search for High-dimensional Continuous Control Space [J]. Computer Science, 2021, 48(10): 30-36. |
[10] | ZHANG Yan-mei, LOU Yin-cheng. Deep Neural Network Based Ponzi Scheme Contract Detection Method [J]. Computer Science, 2021, 48(1): 273-279. |
[11] | LV Ze-yu, LI Ji-xuan, CHEN Ru-Jian and CHEN Dong-ming. Research on Prediction of Re-shopping Behavior of E-commerce Customers [J]. Computer Science, 2020, 47(6A): 424-428. |
[12] | DING Zi-ang, LE Cao-wei, WU Ling-ling and FU Ming-lei. PM2.5 Concentration Prediction Method Based on CEEMD-Pearson and Deep LSTM Hybrid Model [J]. Computer Science, 2020, 47(6A): 444-449. |
[13] | TANG Guo-qiang,GAO Da-qi,RUAN Tong,YE Qi,WANG Qi. Clinical Electronic Medical Record Named Entity Recognition Incorporating Language Model and Attention Mechanism [J]. Computer Science, 2020, 47(3): 211-216. |
[14] | FAN Wei, LIU Ting, HUANG Rui, GUO Qing, ZHANG Bao. Low-level CNN Feature Aided Image Instance Segmentation [J]. Computer Science, 2020, 47(11): 186-191. |
[15] | KONG Fan-yu, ZHOU Yu-feng, CHEN Gang. Traffic Flow Prediction Method Based on Spatio-Temporal Feature Mining [J]. Computer Science, 2019, 46(7): 322-326. |
|