[llvm] [Offload] Allow CUDA Kernels to use arbitrarily large shared memory (PR #145963)
via llvm-commits
llvm-commits at lists.llvm.org
Fri Jun 27 14:34:41 PDT 2025
github-actions[bot] wrote:
<!--LLVM CODE FORMAT COMMENT: {clang-format}-->
:warning: C/C++ code formatter, clang-format found issues in your code. :warning:
<details>
<summary>
You can test this locally with the following command:
</summary>
``````````bash
git-clang-format --diff HEAD~1 HEAD --extensions cpp,h -- offload/plugins-nextgen/cuda/dynamic_cuda/cuda.cpp offload/plugins-nextgen/cuda/dynamic_cuda/cuda.h offload/plugins-nextgen/cuda/src/rtl.cpp
``````````
</details>
<details>
<summary>
View the diff from clang-format here.
</summary>
``````````diff
diff --git a/offload/plugins-nextgen/cuda/src/rtl.cpp b/offload/plugins-nextgen/cuda/src/rtl.cpp
index fd9528061..b899497bd 100644
--- a/offload/plugins-nextgen/cuda/src/rtl.cpp
+++ b/offload/plugins-nextgen/cuda/src/rtl.cpp
@@ -1306,10 +1306,9 @@ Error CUDAKernelTy::launchImpl(GenericDeviceTy &GenericDevice,
// In case we require more memory than the current limit.
if (MaxDynCGroupMem >= MaxDynCGroupMemLimit) {
CUresult AttrResult = cuFuncSetAttribute(
- Func,
- CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES,
- MaxDynCGroupMem);
- Plugin::check(AttrResult,
+ Func, CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES, MaxDynCGroupMem);
+ Plugin::check(
+ AttrResult,
"Error in cuLaunchKernel while setting the memory limits: %s");
MaxDynCGroupMemLimit = MaxDynCGroupMem;
}
``````````
</details>
https://github.com/llvm/llvm-project/pull/145963
More information about the llvm-commits
mailing list