Affiliation:
1. Carnegie Mellon University
2. Nagoya University
Abstract
In this paper, we present two conceptual frameworks for GPU applications to adjust their task execution times based on total workload. These frameworks enable smart GPU resource management when many applications share GPU resources while the workloads of those applications vary. Application developers can explicitly adjust the number of GPU cores depending on their needs. An implicit adjustment will be supported by a run-time framework, which dynamically allocates the number of cores to tasks based on the total workload. The runtime support of the proposed system can be realized using functions which measure the execution times of the tasks on GPU and change the number of GPU cores. We motivate the necessity of this framework in the context of self-driving technologies, and we believe that our frameworks for GPU programming are useful contributions given the increasing emphasis on parallel heterogeneous computing.
Publisher
Association for Computing Machinery (ACM)
Subject
Engineering (miscellaneous),Computer Science (miscellaneous)
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. A Dynamic Protection Mechanism for GPU Memory Overflow;Lecture Notes in Computer Science;2021
2. Hardware accelerator IP cores for real time Radar and camera-based ADAS;Journal of Real-Time Image Processing;2016-12-03
3. CUDA Leaks;ACM Transactions on Embedded Computing Systems;2016-02-20
4. Image Feature Extraction Acceleration;Image Feature Detectors and Descriptors;2016