diff options
author | Michael Kruse <llvm-project@meinersbur.de> | 2021-05-13 17:12:23 -0500 |
---|---|---|
committer | Michael Kruse <llvm-project@meinersbur.de> | 2021-05-13 17:15:34 -0500 |
commit | 83ff0ff46337422171fb36f934bd56c2bc1be15c (patch) | |
tree | 24e7e67bf37944a2ab968a4ca1953e3f0ed4672c /llvm/lib/Bitcode/Reader | |
parent | 93d56922fabaf52eec8d1d4e28e04fa47eb1c797 (diff) | |
download | llvm-83ff0ff46337422171fb36f934bd56c2bc1be15c.zip llvm-83ff0ff46337422171fb36f934bd56c2bc1be15c.tar.gz llvm-83ff0ff46337422171fb36f934bd56c2bc1be15c.tar.bz2 |
[Clang][OpenMP] Allow unified_shared_memory for Pascal-generation GPUs.
The Pascal architecture supports the page migration engine required for
unified_shared_memory, as indicated by NVIDIA:
* https://developer.nvidia.com/blog/unified-memory-cuda-beginners/
* https://developer.nvidia.com/blog/beyond-gpu-memory-limits-unified-memory-pascal/
* https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements
The limitation was introduced in D54493 which justified the cut-off by
the requirement for unified addressing. However, Unified Virtual
Addressing (UVA) is already available with sm20 (Fermi, Kepler,
Maxwell):
* https://docs.nvidia.com/cuda/gpudirect-rdma/index.html#basics-of-uva-cuda-memory-management
Unified shared memory might even be possible with these, but with
migration of entire allocations on kernel startup.
Reviewed By: JonChesterfield
Differential Revision: https://reviews.llvm.org/D101595
Diffstat (limited to 'llvm/lib/Bitcode/Reader')
0 files changed, 0 insertions, 0 deletions