Computer Science ›› 2022, Vol. 49 ›› Issue (3): 86-91.doi: 10.11896/jsjkx.210700199

Special Issue: Big Data & Data Scinece

• Database & Big Data & Data Science • Previous Articles     Next Articles

GSO:A GNN-based Deep Learning Computation Graph Substitutions Optimization Framework

MIAO Xu-peng1, ZHOU Yue1, SHAO Ying-xia2, CUI Bin1   

  1. 1 School of Electronics Engineering and Computer Science,Peking University,Beijing 100871,China
    2 School of Computer Science,Beijing University of Posts and Telecommunications,Beijing 100871,China
  • Received:2021-07-19 Revised:2021-08-16 Online:2022-03-15 Published:2022-03-15
  • About author:MIAO Xu-peng,born in 1995,Ph.D candidate,is a student member of China Computer Federation.His main research interests include deep learning system and distributed optimization.
    CUI Bin,born in 1975,Ph.D,professor,Ph.D supervisor,is a distinguished member of China Computer Federation.His main research interests include database system architectures,query and index techniques,big data management and mining.
  • Supported by:
    National Key Research and Development Program of China(2018YFB1004403),National Natural Science Foundation of China(61832001) and PKU-Tencent Joint Research Lab.

Abstract: Deep learning has achieved great success in various practical applications.How to effectively improve the model execution efficiency is one of the important research issues in this field.The existing deep learning frameworks usually model deep learning in the form of computational graphs,try to optimize computational graphs through subgraph substitution rules designed by experts and mainly use heuristic algorithms to search substitution sequences.Their shortcomings mainly include:1)the exis-ting subgraph substitution rules result in a large search space and the heuristic algorithms are not efficient;2)these algorithms are not scalable for large computation graphs;3)cannot utilize the history optimization results.In order to solve the above problem,we propose GSO,a graph neural network-based deep learning computation graph optimization framework.We transfer the graph substitution optimization problem as the subgraph matching problem.Based on the feature information from the operators and the computation graph topology,we utilize the graph neural network to predict the subgraph matching feasibility and positions.We implement the framework using Python,which is compatible with the mainstream deep learning systems.The experimental results show that:1)compared to the total graph substitution rules,the proposed rule can reduce the search space by up to 92%;2)compared to the existing heuristic algorithms,GSO can complete the subgraph replacement process of the computational graph 2 times faster.The optimized computation graph is up to 34% faster the original graph.

Key words: Deep learning, Graph neural network, Graph substitutions, Optimizing DNN computation

CLC Number: 

  • TP311
[1]CANZIANI A,PASZKE A,CULURCIELLO E.An analysis ofdeep neural network models for practical applications[J].arXiv:1605.07678v4,2016.
[2]MARTÍN A,PAUL B,JIANMIN C,et al.TensorFlow:a sys-tem for large-scale machine learning[C]//12th USENIX Symposium on Operating Systems Design and Implementation.2016:265-283.
[3]JIA Z H,JAMES T,TODD W,et al.Optimizing DNN Computation with Relaxed Graph Substitutions[C]//Proceedings of the 2nd SysML Conference.2019.
[4]JIA Z H,ODED P,JAMES T,et al.TASO:optimizing deeplearning computation with automatic generation of graph substitutions[C]//Proceedings of the 27th ACM Symposium on Operating Systems Principles.2019:47-62.
[5]FANG J Z,SHEN Y Y,WANG Y,et al.Optimizing DNN computation graph using graph substitutions[C]//Proceedings of the VLDB Endowment.2020:2734-2746.
[6]CHEN T Q,THIERRY M,JIANG Z H,et al.TVM:an automated end-to-end optimizing compiler for deep learning[C]//13th USENIX Symposium on Operating Systems Design and Implementation.2018:578-594.
[7]PASZKE A,GROSS S,MASSA F,et al.Pytorch:An imperative style,high-performance deep learning library[C]//Advances in Neural Information Processing Systems.2019,32:8026-8037.
[8]Nvidia tensorrt:Programmable inference accelerator[Z/OL].https://developer.nvidia.com/tensorrt.
[9]THOMAS N,MAX W.Semi-supervised classification withgraph convolutional networks[J].arXiv:1609.02907,2016.
[10]MIAO X P,ZHANG W T,SHAO Y X,et al.Lasagne:A multi-layer graph convolutional network framework via node-aware deep architecture[J/OL].IEEE Transactions on Knowledge and Data Engineering,2021.https://ieeexplore.ieee.org/document/9513581.
[11]MIAO X P,GÜREL N M,ZHANG W T,et al.Degnn:Improving graph neural networks with graph decomposition[C]//Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining.2021:1223-1233.
[12]ZHANG W T,MIAO X P,SHAO Y X,et al.Reliable Data Distillation on Graph Convolutional Network [C]//Proceedings of the 2020 ACM SIGMOD International Conference on Management of Data.2020:1399-1414.
[13]MINAR M R,NAHER J.Recent advances in deep learning:An overview[J].arXiv:1807.08169,2018.
[14]MIAO X P,MA L X,YANG Z,et al.Cuwide:towards efficient flow-based training for sparse wide models on gpus[J/OL].IEEE Transactions on Knowledge and Data Engineering,2020.https://ieeexplore.ieee.org/document/9261124.
[15]HE K,ZHANG X Y,REN S Q,et al.Deep Residual Learning for Image Recognition [C]//Proceedings of the IEEE Confe-rence on Computer Vision and Pattern Recognition.2016:770-778.
[16]ALEX K,ILYA S,GEOFFREY E.ImageNet classification with deep convolutional neural networks [C]//Advances in Neural Information Processing Systems.2012,25:1097-1105.
[17]KAREN S,ANDREW Z.Very deep convolutional networks for large-scale image recognition[J].arXiv:1409.1556v4,2014.
[18]ZAREMBA W,ILYA S,ORIOL V.Recurrent neural network regularization[J].arXiv:1409.2329,2014.
[1] ZHOU Fang-quan, CHENG Wei-qing. Sequence Recommendation Based on Global Enhanced Graph Neural Network [J]. Computer Science, 2022, 49(9): 55-63.
[2] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[3] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[4] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[5] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[6] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[7] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[8] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[9] YAN Jia-dan, JIA Cai-yan. Text Classification Method Based on Information Fusion of Dual-graph Neural Network [J]. Computer Science, 2022, 49(8): 230-236.
[10] QI Xiu-xiu, WANG Jia-hao, LI Wen-xiong, ZHOU Fan. Fusion Algorithm for Matrix Completion Prediction Based on Probabilistic Meta-learning [J]. Computer Science, 2022, 49(7): 18-24.
[11] YANG Bing-xin, GUO Yan-rong, HAO Shi-jie, Hong Ri-chang. Application of Graph Neural Network Based on Data Augmentation and Model Ensemble in Depression Recognition [J]. Computer Science, 2022, 49(7): 57-63.
[12] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[13] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[14] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[15] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!