-
Notifications
You must be signed in to change notification settings - Fork 906
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEA] Add libcudf any() and all() reductions #1874
Comments
@kovaltan FYI in case you have thoughts on this. |
I agree to have
The current implementation of reduction has already output dtype option. Btw, I've found a bug about |
In my experience even CUB doesn't have an optimized 1-bit-per-boolean reduction (i.e. using |
@harrism Why not use thrust::all_of and thrust::any_of ? |
|
faster early out for |
@karthikeyann Your results in that issue are fishy/impossible. |
Yes. looked into it. it had a bug. updating it now. |
Is your feature request related to a problem? Please describe.
Currently cuDF Python implements any() and all() for a column by calling libcudf's min() and max() (see NVIDIA/thrust#1863). libcudf should provide any() and all() instead. They can be made faster and also give us better control over semantics.
Describe the solution you'd like
It would be faster to provide an optimized any() and all() that first check that the values are non-zero to get a boolean per element and then use bitwise reductions in CUDA for a faster reduction.
Describe alternatives you've considered
It currently works, so this is just a potential optimization.
The text was updated successfully, but these errors were encountered: