File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

postgraduate thesis: Fine-grained concurrent kernel execution on SM/CU level for GPGPU computing

TitleFine-grained concurrent kernel execution on SM/CU level for GPGPU computing
Authors
Advisors
Advisor(s):Wang, CL
Issue Date2019
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Wu, H. [吴昊]. (2019). Fine-grained concurrent kernel execution on SM/CU level for GPGPU computing. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractAs a critical computing resource in multi-user systems such as supercomputers, data centres and cloud services, a GPU contains multiple compute units (CUs). GPU Multitasking is an intuitive solution to underutilization in GPGPU computing. Recently proposed solutions of multitasking GPUs can be classified into two categories: (1) spatially-partitioned sharing (SPS), which co-executes different kernels on disjointed sets of compute units (CU), and (2) simultaneous multikernel (SMK), which runs multiple kernels simultaneously within a CU. Compared to SPS, SMK can improve resource utilization even further due to the interleaving of instructions from kernels with low dynamic resource contentions. However, it is hard to implement SMK on current GPU architecture, because: (1) techniques for applying SMK on top of GPU hardware scheduling policy are scarce; (2) finding an efficient SMK scheme is difficult due to the complex interferences of concurrently executed kernels. In this paper, we propose a lightweight and effective performance model to evaluate the complex interferences of SMK. Based on the probability of independent events, our performance model is built from a totally new angle and contains limited parameters. Then, we propose a metric, symbiotic factor, which can evaluate an SMK scheme so that kernels with complementary resource utilization can co-run within a CU. Also, we analyse the advantages and disadvantages of kernel slicing and kernel stretching techniques and integrate them to apply SMK on GPUs instead of simulators. We validate our model on 18 benchmarks. Compared to the optimized hardware-based concurrent kernel execution whose kernel launching order brings fast execution time, the results of co-running kernel pairs show 11%, 18% and 12% speedup on AMD R9 290X, RX 480 and Vega 64, on average. Compared to the Warped-Slicer, the results show 29%, 18% and 51% speedup on AMD R9 290X, RX 480 and Vega 64, on average.
DegreeDoctor of Philosophy
SubjectGraphics processing units - Programming
Computer graphics
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/281608

 

DC FieldValueLanguage
dc.contributor.advisorWang, CL-
dc.contributor.authorWu, Hao-
dc.contributor.author吴昊-
dc.date.accessioned2020-03-18T11:33:04Z-
dc.date.available2020-03-18T11:33:04Z-
dc.date.issued2019-
dc.identifier.citationWu, H. [吴昊]. (2019). Fine-grained concurrent kernel execution on SM/CU level for GPGPU computing. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/281608-
dc.description.abstractAs a critical computing resource in multi-user systems such as supercomputers, data centres and cloud services, a GPU contains multiple compute units (CUs). GPU Multitasking is an intuitive solution to underutilization in GPGPU computing. Recently proposed solutions of multitasking GPUs can be classified into two categories: (1) spatially-partitioned sharing (SPS), which co-executes different kernels on disjointed sets of compute units (CU), and (2) simultaneous multikernel (SMK), which runs multiple kernels simultaneously within a CU. Compared to SPS, SMK can improve resource utilization even further due to the interleaving of instructions from kernels with low dynamic resource contentions. However, it is hard to implement SMK on current GPU architecture, because: (1) techniques for applying SMK on top of GPU hardware scheduling policy are scarce; (2) finding an efficient SMK scheme is difficult due to the complex interferences of concurrently executed kernels. In this paper, we propose a lightweight and effective performance model to evaluate the complex interferences of SMK. Based on the probability of independent events, our performance model is built from a totally new angle and contains limited parameters. Then, we propose a metric, symbiotic factor, which can evaluate an SMK scheme so that kernels with complementary resource utilization can co-run within a CU. Also, we analyse the advantages and disadvantages of kernel slicing and kernel stretching techniques and integrate them to apply SMK on GPUs instead of simulators. We validate our model on 18 benchmarks. Compared to the optimized hardware-based concurrent kernel execution whose kernel launching order brings fast execution time, the results of co-running kernel pairs show 11%, 18% and 12% speedup on AMD R9 290X, RX 480 and Vega 64, on average. Compared to the Warped-Slicer, the results show 29%, 18% and 51% speedup on AMD R9 290X, RX 480 and Vega 64, on average.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshGraphics processing units - Programming-
dc.subject.lcshComputer graphics-
dc.titleFine-grained concurrent kernel execution on SM/CU level for GPGPU computing-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.identifier.doi10.5353/th_991044214993603414-
dc.date.hkucongregation2020-
dc.identifier.mmsid991044214993603414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats