DocumentCode :
1996730
Title :
Dynamic Sharing of GPUs in Cloud Systems
Author :
Diab, Khaled M. ; Rafique, M. Mustafa ; Hefeeda, Mohamed
Author_Institution :
Qatar Comput. Res. Inst. (QCRI), Qatar Found., Doha, Qatar
fYear :
2013
fDate :
20-24 May 2013
Firstpage :
947
Lastpage :
954
Abstract :
The use of computational accelerators, specifically programmable GPUs, is becoming popular in cloud computing environments. Cloud vendors currently provide GPUs as dedicated resources to cloud users, which may result in under-utilization of the expensive GPU resources. In this work, we propose gCloud, a framework to provide GPUs as on-demand computing resources to cloud users. gCloud allows on-demand access to local and remote GPUs to cloud users only when the target GPU kernel is ready for execution. In order to improve the utilization of GPUs, gCloud efficiently shares the GPU resources among concurrent applications from different cloud users. Moreover, it reduces the inter-application interference of concurrent kernels for GPU resources by considering the local and global memory, number of threads, and the number of thread blocks of each kernel. It schedules concurrent kernels on available GPUs such that the overall inter-application interference across the cluster is minimal. We implemented gCloud as an independent module, and integrated it with the Open Stack cloud computing platform. Evaluation of gCloud using representative applications shows that it improves the utilization of GPU resources by 56.3% on average compared to the current state-of-the-art systems that serialize GPU kernel executions. Moreover, gCloud significantly reduces the completion time of GPU applications, e.g., in our experiments of running a mix of 8 to 28 GPU applications on 4 NVIDIA Tesla GPUs, gCloud achieves up to 430% reduction in the total completion time.
Keywords :
cloud computing; graphics processing units; Cloud Systems; Dynamic Sharing; GPU kernel executions; cloud computing environments; computational accelerators; concurrent applications; expensive GPU resources; gCloud; global memory; inter-application interference; local memory; on-demand computing resources; open stack cloud computing platform; Cloud computing; Context; Graphics processing units; Instruction sets; Kernel; Memory management; Message systems;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Parallel and Distributed Processing Symposium Workshops & PhD Forum (IPDPSW), 2013 IEEE 27th International
Conference_Location :
Cambridge, MA
Print_ISBN :
978-0-7695-4979-8
Type :
conf
DOI :
10.1109/IPDPSW.2013.102
Filename :
6650978
Link To Document :
بازگشت