Efficient GPU Offloading with OpenMP for a Hyperbolic Finite Volume Solver on Dynamically Adaptive Meshes
Monday, May 22, 2023 4:50 PM to 5:15 PM · 25 min. (Europe/Berlin)
Hall 4 - Ground Floor
Research Paper
Emerging HPC Processors and AcceleratorsExascale SystemsMemory and Storage TechnologyParallel Programming LanguagesPerformance Modeling and Tuning
Information
We identify and show how to overcome an OpenMP bottleneck in the administration of GPU memory.
It arises for a wave equation solver on dynamically adaptive block-structured Cartesian meshes, which keeps all CPU threads busy and allows all of them to offload sets of patches to the GPU.
Our studies show that multithreaded, concurrent, non-deterministic access to the GPU leads to performance breakdowns, since the GPU memory bookkeeping as offered through OpenMP's map clause, i.e., the allocation and freeing, becomes another runtime challenge besides expensive data transfer and actual computation. We, therefore, propose to retain the memory management responsibility on the host: A caching mechanism acquires memory on the accelerator for all CPU threads, keeps hold of this memory and hands it out to the offloading threads upon demand.
We show that this user-managed, CPU-based memory administration helps us to overcome the GPU memory bookkeeping bottleneck and speeds up the time-to-solution of Finite Volume kernels by more than an order of magnitude.
Contributors:
Contributors:
Format
On-siteOn Demand
Beginner Level
40%
Intermediate Level
30%
Advanced Level
30%