You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I find this library when looking for a trivially copyable tuple implementation in my CUDA code. According to the CUDA C Programming Guide https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#global-function-argument-processing the argument passed to a CUDA kernel must be trivially copyable, so std::tuple is not a choice. However, I find this library do not support CUDA in lack of __device__ modifiers. Would you consider add CUDA support to this library? Specifically, when __CUDACC__ macro is defined, add __host__ __device__ to each of the member functions?
The text was updated successfully, but these errors were encountered:
Hi, I find this library when looking for a trivially copyable tuple implementation in my CUDA code. According to the CUDA C Programming Guide https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#global-function-argument-processing the argument passed to a CUDA kernel must be trivially copyable, so std::tuple is not a choice. However, I find this library do not support CUDA in lack of
__device__
modifiers. Would you consider add CUDA support to this library? Specifically, when__CUDACC__
macro is defined, add__host__ __device__
to each of the member functions?The text was updated successfully, but these errors were encountered: