Computer Science ›› 2024, Vol. 51 ›› Issue (7): 22-28.doi: 10.11896/jsjkx.230500220

• Computer Software • Previous Articles     Next Articles

Study on Deep Learning Automatic Scheduling Optimization Based on Feature Importance

YANG Heng1,2, LIU Qinrang2, FAN Wang2, PEI Xue2, WEI Shuai2, WANG Xuan1,2   

  1. 1 College of Cyberspace Security,Zhengzhou University,Zhengzhou 450003,China
    2 Institute of Information Technology,Information Engineering University,Zhengzhou 450002,China
  • Received:2023-05-30 Revised:2023-10-13 Online:2024-07-15 Published:2024-07-10
  • About author:YANG Heng,born in 1998,postgra-duate.His main research interest is deep learning compiler.
    LIU Qinrang,born in 1975,Ph.D,professor,Ph.D supervisor.His main research interests include cyberspace security and chip design.
  • Supported by:
    Major Project of National Key R & D Program of China(2022YFB4401401) and Program of Songshan Laboratory(included in the management of Major Science and Technology Program of Henan Province)(221100211100-01).

Abstract: With the rapid development of deep learning and hardware architectures,the diversity of models and hardware architectures make the deployment for deep learning models with high performance manually become increasingly challenging.So current AI compiler framework often adopts automatic scheduling.Since the existing optimization to TVM automatic scheduling has such issues as unbalanced data sets in cost model and overlong scheduling time,an automatic scheduling optimization strategy based on feature importance is designed in this paper.First,the feature importance is analyzed through the xgboost algorithm.Then a stra-tegy that reduce the data feature dimensions based on the importance coefficient and reassign the data labels is adopted to improve the precision of the cost model and optimize the efficiency of the automatic scheduling.Experiment results show that the proposed optimization method can reduce the automatic scheduling time of three kinds of deep learning models by 9.7%~17.2%,and reduce the inference time by up to 15%.

Key words: AI compiler, Automatic scheduling, xgboost, Feature importance, Deep learning

CLC Number: 

  • TP302
[1]JIA Y,SHELHAMER E,DONAHUE J,et al.Caffe:Convolutional architecture for fast feature embedding[C]//Proceedings of the 22nd ACM International Conference on Multimedia.NewYork:Association for Computing Machinery,2014:675-678.
[2]ABADI M,BARHAM P,CHEN J,et al.Tensorflow:A system for large-scale machine learning[C]//12th{USENIX} Sympo-sium on Operating Systems Design and Implementation({Osdi}16).Sacannah,GA,USA:USENIX Association,2016:265-283.
[3]PASZKE A,GROSS S,MASSA F,et al.PyTorch:An Imperative Style,High-Performance Deep Learning Library[C]//Advances in Neural Information Processing Systems 32(NeurIPS 2019).Vancouver,Canada,2019:8024-8035.
[4]CHEN T,LI M,LI Y,et al.MXNet:A Flexible and EfficientMachine Learning Library for Heterogeneous Distributed Systems[J].arXiv:1512.01274,2015.
[5]CHETLUR S,WOOLLEY C,VANDERMERSCH P,et al.cu-DNN:Efficient Primitives for Deep Learning[J].arXiv:1410.0759,2014.
[6]NVIDIA.TensorRT Github repository[EB/OL].[2020-02-04].https://github.com/NVIDIA/TensorRT.
[7]GAO J,LIU S,HUANG Z Q,et al.Deep Neural Network Ope-rator Acceleration Library Optimization Based on Domestic Many-core Processor [J].Computer Science,2022,49(5):355-362.
[8]CHEN T,MOREAU T,JIANG Z,et al.{TVM}:An automated end-to-end optimizing compiler for deep learning[C]//13th{USENIX} Symposium on Operating Systems Design and Implementation({OSDI}18).Berkeley:{USENIX}Association,2018:578-594.
[9]ZHENG S,LIANG Y,WANG S,et al.FlexTensor:An Auto-matic Schedule Exploration and Optimization Framework for Tensor Computation on Heterogeneous System[C]//ASPLOS'20:Architectural Support for Programming Languages and Operating Systems.NewYork:Association for Computing Machi-nery,2020:859-873.
[10]ROTEM N,FIX J,ABDULRASOOL S,et al.Glow:GraphLowering Compiler Techniques for Neural Networks[J].ar-Xiv:1805.00907,2018.
[11]CYPHERS S,BANSAL A K,BHIWANDIWALLA A,et al.Intel nGraph:An Intermediate Representation,Compiler,and Executor for Deep Learning[J].arXiv:1801.08058,2018.
[12]CHEN T,ZHENG L,YAN E,et al.Learning to Optimize Tensor Programs[J].arXiv:1805.08166,2018.
[13]ZHENG L,JIA C,SUN M,et al.Ansor:Generating High-Performance Tensor Programs for Deep Learning[J].arXiv:2006.06762,2020.
[14]ROESCH J,LYUBOMIRSKY S,KIRISAME M,et al.Relay:AHigh-Level IR for Deep Learning[J].arXiv:1904.08368,2019.
[15]ROESCH J,LYUBOMIRSKY S,WEBER L,et al.Relay:a new IR for machine learning frameworks[C]//Proceedings of the 2nd ACM SIGPLAN International Workshop on Machine Learning and Programming Languages(MAPL 2018).New York:Association for Computing Machinery,2018:58-68.
[16]VIKHAR P.A Evolutionary algorithms:A critical review andits future prospects[C]//International Conference on Global Trends in Signal Processing,Information Computing and Communication(ICGTSPICC).IEEE,2016:261-265.
[17]LIU G H,LI Y,WANG X L.Optimization of Deep Learning Compiler Acceleration Technology for Aerospace Heterogeneous Platforms[J].Aerospace Control,2022,40(2):60-65.
[18]CHEN T,GUESTRIN C.Xgboost:A scalable tree boosting system[C]//Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining(KDD'16).New York:Association for Computing Machinery,2016:785-794.
[19]ZHAO J Q.Research on Compiler auto-tuning Method Based on Deep Reinforce Learning[D].Xi'an:Northwest University,2022.
[20]RYU J,SUNG H.MetaTune:Meta-Learning Based Cost Model for Fast and Efficient Auto-tuning Frameworks[J].arXiv:2102.04199,2021.
[21]MU J,WANG M,LI L,et al.A history-based auto-tuningframework for fast and high-performance DNN design on GPU[C]//57th ACM/IEEE Design Automation Conference(DAC).IEEE Press,2020:1-6.
[22]ZHENG L,LIU R,SHAO J,et al.TenSet:A Large-scale Program Performance Dataset for Learned Tensor Compilers[C]//Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track(Round 1).2021.
[1] LI Jiaying, LIANG Yudong, LI Shaoji, ZHANG Kunpeng, ZHANG Chao. Study on Algorithm of Depth Image Super-resolution Guided by High-frequency Information ofColor Images [J]. Computer Science, 2024, 51(7): 197-205.
[2] SHI Dianxi, GAO Yunqi, SONG Linna, LIU Zhe, ZHOU Chenlei, CHEN Ying. Deep-Init:Non Joint Initialization Method for Visual Inertial Odometry Based on Deep Learning [J]. Computer Science, 2024, 51(7): 327-336.
[3] FAN Yi, HU Tao, YI Peng. Host Anomaly Detection Framework Based on Multifaceted Information Fusion of SemanticFeatures for System Calls [J]. Computer Science, 2024, 51(7): 380-388.
[4] GAN Run, WEI Xianglin, WANG Chao, WANG Bin, WANG Min, FAN Jianhua. Backdoor Attack Method in Autoencoder End-to-End Communication System [J]. Computer Science, 2024, 51(7): 413-421.
[5] WANG Yingjie, ZHANG Chengye, BAI Fengbo, WANG Zumin. Named Entity Recognition Approach of Judicial Documents Based on Transformer [J]. Computer Science, 2024, 51(6A): 230500164-9.
[6] LIANG Fang, XU Xuyao, ZHAO Kailong, ZHAO Xuanfeng, ZHANG Guijun. Remote Template Detection Algorithm and Its Application in Protein Structure Prediction [J]. Computer Science, 2024, 51(6A): 230600225-7.
[7] PENG Bo, LI Yaodong, GONG Xianfu, LI Hao. Method for Entity Relation Extraction Based on Heterogeneous Graph Neural Networks and TextSemantic Enhancement [J]. Computer Science, 2024, 51(6A): 230700071-5.
[8] ZHANG Tianchi, LIU Yuxuan. Research Progress of Underwater Image Processing Based on Deep Learning [J]. Computer Science, 2024, 51(6A): 230400107-12.
[9] WANG Guogang, DONG Zhihao. Lightweight Image Semantic Segmentation Based on Attention Mechanism and Densely AdjacentPrediction [J]. Computer Science, 2024, 51(6A): 230300204-8.
[10] WANG Li, CHEN Gang, XIA Mingshan, HU Hao. DUWe:Dynamic Unknown Word Embedding Approach for Web Anomaly Detection [J]. Computer Science, 2024, 51(6A): 230300191-5.
[11] HUANG Haixin, CAI Mingqi, WANG Yuyao. Review of Point Cloud Semantic Segmentation Based on Graph Convolutional Neural Networks [J]. Computer Science, 2024, 51(6A): 230400196-7.
[12] LYU Yiming, WANG Jiyang. Iron Ore Image Classification Method Based on Improved Efficientnetv2 [J]. Computer Science, 2024, 51(6A): 230600212-6.
[13] YANG Xiuzhang, WU Shuai, REN Tianshu, LIAO Wenjing, XIANG Meiyu, YU Xiaomin, LIU Jianyi, CHEN Dengjian. Complex Environment License Plate Recognition Algorithm Based on Improved Image Enhancement and CNN [J]. Computer Science, 2024, 51(6A): 220200162-7.
[14] SONG Zhen, WANG Jiqiang, HOU Moyu, ZHAO Lin. Conveyor Belt Defect Detection Network Combining Attention Mechanism with Line Laser Assistance [J]. Computer Science, 2024, 51(6A): 230800115-6.
[15] WU Chunming, LIU Yali. Method for Lung Nodule Detection on CT Images Using Improved YOLOv5 [J]. Computer Science, 2024, 51(6A): 230500019-6.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!