Tehnički glasnik, Vol. 19 No. 3, 2025.
Prethodno priopćenje
https://doi.org/10.31803/tg-20240112104444
Efficient Deep Learning Job Allocation in Cloud Systems by Predicting Resource Consumptions including GPU and CPU
Abuda Chad Ferrino
orcid.org/0009-0002-4448-6287
; Department of Computer AI Convergence Engineering, Kumoh National Institute of Technology, 61 Daehak-ro, Gumi-si, Gyeongsangbuk-do, 39177, Republic of Korea
Tae Young Choe
; Department of Computer AI Convergence Engineering, Kumoh National Institute of Technology, 61 Daehak-ro, Gumi-si, Gyeongsangbuk-do, 39177, Republic of Korea
*
* Dopisni autor.
Sažetak
One objective of GPU scheduling in cloud systems is to minimize the completion times of given deep learning models. This is important for deep learning in cloud environments because deep learning workloads require a lot of time to finish, and misallocation of these workloads can create a huge increase in job completion time. Difficulties of GPU scheduling come from a diverse type of parameters including model architectures and GPU types. Some of these model architectures are CPU-intensive rather than GPU-intensive which creates a different hardware requirement when training different models. The previous GPU scheduling research had used a small set of parameters, which did not include CPU parameters, which made it difficult to reduce the job completion time (JCT). This paper introduces an improved GPU scheduling approach that reduces job completion time by predicting execution time and various resource consumption parameters including GPU Utilization%, GPU Memory Utilization%, GPU Memory, and CPU Utilization%. The experimental results show that the proposed model improves JCT by up to 40.9% on GPU Allocation based on Computing Efficiency compared to Driple.
Ključne riječi
cloud computing; convolutional neural network; deep learning; GPU job scheduling; performance estimation
Hrčak ID:
332174
URI
Datum izdavanja:
15.9.2025.
Posjeta: 156 *