Computer Science ›› 2025, Vol. 52 ›› Issue (11A): 241000036-5.doi: 10.11896/jsjkx.241000036

• Big Data & Data Science • Previous Articles     Next Articles

Deep Neural Network-based Resource Allocation for Large-scale Operation Simulation

YE Shuai, LI Hao, SHI Peiteng, HUANG Yulin   

  1. Academy of Military Science,Beijing 100091,China
  • Online:2025-11-15 Published:2025-11-10
  • Supported by:
    Young Scientists Fund of the National Natural Science Foundation of China(62102446).

Abstract: With the development of artificial intelligence,operation experiments tend to be intelligent.Large-scale operation simulation is an important support for conducting intelligent operation experiments and an effective means to solve problems such as multiple variables and complex combinations in operation experiments.It has the characteristics of large sample size and high speed requirements.The high-speed operation of massive simulation samples depends on the efficient scheduling of high-perfor-mance hardware clusters,which faces the problems of large differences in computing resource requirements and difficult manual allocation.How to accurately predict and dynamically allocate the resources required for each sample is the key to improving the efficiency of large-scale simulation.This paper proposes a deep neural network(DNN)-based resource prediction model for large-scale operation simulation.The method firstly constructs a deep neural network in-loop simulation resource management architecture.Secondly,it constructs a deep neural network prediction model by extracting features and learning from combat simulation sample files.During the operation of large-scale simulation,it achieves accurate prediction and dynamic allocation of massive ope-ration simulation job resources by online predicting the computing resources required for each sample.Test results show that in a typical operation experiment simulation scenario with thousands of samples,theproposed prediction model reduces the completion time by 20.8% on 10 high-performance server nodes compared to traditional configuration methods.

Key words: Deep neural network, Large-scale simulation, Resource prediction, Cluster management

CLC Number: 

  • TP391.9
[1]LI G J.AI4R:The fifth scientific research paradigm[J].Bulletin of Chinese Academy of Sciences,2024,39(1):1-9.
[2]WANG J,CHEN R,TANG L.Research on the technology of Army intelligent combat exercise with large sample[J].National Defense Science and Technology,2020,41(1):41-44.
[3]WU X P,CAO D,DONG S,et al.Large-sample simulation design and verification of system simulation platform[J/OL].Computer Measurement and Control,1-8[2024-05-17].http://kns.cnki.net/kcms/detail/11.4762.TP.20240221.2108.006.html.
[4]LU W R.Research and improvement of container cloud resource scheduling strategy based onKubernetes[D].Hangzhou:Zhejiang Sci-Tech University,2023.
[5]JASAK,ALEKSANDAR J H,TUKOVIĆ,et al.OpenFOAM:A C++ Library for Complex Physics Simulations[C]//Coupled Methods in Numerical Dyamics.2007.
[6]ZHAO Z,et al.Design of general CFD software PHengLEI [J].Computer Engineering & Science,2020,42(2):210-219.
[7]LIU H,DONG X Y,YANG Z H.BiGRU-LGB cloud load prediction model based on stacking framework[J].Journal of Xidian University,2023,50(3):83-94,104.
[8]THINAKARAN P,GUNASEKARAN J R,SHARMA B,et al.Kube-Knots:Resource Harvesting through Dynamic Container Orchestration in GPU-based Datacenters[C]//2019 IEEE International Conference on Cluster Computing(CLUSTER).2019:1-13.
[9]TANG S,LEE B S,HE B.Fair Resource Allocation for Data-Intensive Computing in the Cloud[J].IEEE Transactions on Ser-vices Computing,2018,11(1):20-33.
[10]CHEN Z,HU J,MIN G,et al.Adaptive and Efficient Resource Allocation in Cloud Datacenters Using Actor-Critic Deep Reinforcement Learning[C]//IEEE Transactions on Parallel and Distributed Systems.2022:1911-1923.
[11]HU Z,LI D,ZHANG D,et al.ReLoca:Optimize Resource Al-location for Data-parallel Jobs using Deep Learning[C]//IEEE INFOCOM 2020-IEEE Conference on Computer Communications.2020:1163-1171.
[12]CHEN S Y,ZHUANG Y,LI J.Multiple load prediction model for mobile cloud computing based on LSTM network[J].Computer and Modernization,2021(6):74-85.
[13]LIANG R H,XIE X L,ZHAI Q H,et al.Research on container cloud load prediction based on improved stacking ensemble mo-del[J].Computer Applications & Software,2023,40(12):48-55,100.
[14]PASZKE A,GROSS S,MASSA F,et al.Pytorch:An imperative style,high-performance deep learning library[C]//Advances in Neural Information Processing Systems.2019.
[1] XIA Zhuoqun, ZHOU Zihao, DENG Bin, KANG Chen. Security Situation Assessment Method for Intelligent Water Resources Network Based on ImprovedD-S Evidence [J]. Computer Science, 2025, 52(6A): 240600051-6.
[2] LI Zhoucheng, ZHANG Yi, SUN Jin. Stochastic Optimization Method for Multi-exit Deep Neural Networks for Edge Intelligence Applications [J]. Computer Science, 2025, 52(4): 85-93.
[3] LIN Zheng, LIU Sicong, GUO Bin, DING Yasan, YU Zhiwen. Adaptive Operator Parallel Partitioning Method for Heterogeneous Embedded Chips in AIoT [J]. Computer Science, 2025, 52(2): 299-309.
[4] CHEN Ping’an, DENG Qi. Expression Detection Algorithm Based on SSD Network Model Reconstruction [J]. Computer Science, 2025, 52(11A): 250200066-6.
[5] HUANG Xinli, GAO Guoju. Adaptive Gradient Sparsification Approach to Training Deep Neural Networks [J]. Computer Science, 2025, 52(11A): 250100106-6.
[6] WANG Liuyi, ZHOU Chun, ZENG Wenqiang, HE Xingxing, MENG Hua. High-frequency Feature Masking-based Adversarial Attack Algorithm [J]. Computer Science, 2025, 52(10): 374-381.
[7] ZHU Fukun, TENG Zhen, SHAO Wenze, GE Qi, SUN Yubao. Semantic-guided Neural Network Critical Data Routing Path [J]. Computer Science, 2024, 51(9): 155-161.
[8] HAN Bing, DENG Lixiang, ZHENG Yi, REN Shuang. Survey of 3D Point Clouds Upsampling Methods [J]. Computer Science, 2024, 51(7): 167-196.
[9] XU Xiaohua, ZHOU Zhangbing, HU Zhongxu, LIN Shixun, YU Zhenjie. Lightweight Deep Neural Network Models for Edge Intelligence:A Survey [J]. Computer Science, 2024, 51(7): 257-271.
[10] ZHU Jin, TAO Chuanqi, GUO Hongjing. Test Input Prioritization Approach Based on DNN Model Output Differences [J]. Computer Science, 2024, 51(6A): 230600121-8.
[11] LI Wenting, XIAO Rong, YANG Xiao. Improving Transferability of Adversarial Samples Through Laplacian Smoothing Gradient [J]. Computer Science, 2024, 51(6A): 230800025-6.
[12] ZHONG Zhenyu, LIN Yongliang, WANG Haotian, LI Dongwen, SUN Yufei, ZHANG Yuzhi. Automatic Pipeline Parallel Training Framework for General-purpose Computing Devices [J]. Computer Science, 2024, 51(12): 129-136.
[13] ZHAO Ruonan, LI Duo, SONG Jiangling, ZHANG Rui. Automatic Sleep Staging Based on Multimodal Data and Fusion Deep Network [J]. Computer Science, 2024, 51(11A): 231100160-6.
[14] ZHANG Mingze, LI Yi, WU Wenyuan, SHI Mingquan, WANG Zhengjiang. FCTNet:Bus Arrival Time Prediction Method Based on Dual Domain Deep Learning [J]. Computer Science, 2024, 51(11A): 231000180-7.
[15] WANG Huaiqin, LUO Jian, WANG Haiyan. Feature Weight Perception-based Prediction of Virtual Network Function Resource Demands [J]. Computer Science, 2023, 50(9): 331-336.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!