Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cublaslt ran into an error! On h100 #599

Closed
Ar770 opened this issue Jul 17, 2023 · 11 comments
Closed

cublaslt ran into an error! On h100 #599

Ar770 opened this issue Jul 17, 2023 · 11 comments
Labels
bug Something isn't working enhancement New feature or request high priority (first issues that will be worked on)

Comments

@Ar770
Copy link

Ar770 commented Jul 17, 2023

I'm not sure what is the status of the other issues opened related to this problem, I can confirm that the problem remaines with 0.40.1 on an H100 instance.
See here the complete error:
#538

I was in touch with the Lambdalabs support, maybe it could help.
I added their response here:
#383

@TimDettmers
Copy link
Collaborator

Thanks for letting me know. I looked at the details and it turns out Hopper does not support col_turning or col_ampere formats. As such, new logic is needed. I might implement the FP8 version instead for Hopper.

@TimDettmers TimDettmers added bug Something isn't working enhancement New feature or request high priority (first issues that will be worked on) labels Jul 17, 2023
@Ar770
Copy link
Author

Ar770 commented Jul 17, 2023

Thank you!
Do you have ETA to resolve this issue?

@Mythli
Copy link

Mythli commented Aug 9, 2023

Hi @TimDettmers ,

Thank you for your amazing work.
While I'm not qualified to fix this issue myself I can provide a bounty so you could focus on it.

I can provide an H100 for testing as well.

Please reach out to me at [email protected]

@jieWANGforwork
Copy link

I encountered the same error as loading 8bit Llama on H100. Are there any solutions?

Copy link

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

@hyaticua
Copy link

hyaticua commented Apr 2, 2024

Hi this still seems to be an issue.

@corbt
Copy link

corbt commented Apr 15, 2024

Confirming this is still an issue, and would love to get it fixed so we can use H100s for training!

@karroo
Copy link

karroo commented Apr 23, 2024

I have the same error on H100!!!

@zhangj1an
Copy link

still having the same issue

@Jean-KOUAGOU
Copy link

I get the following: Error operation not supported at line 380 in file /mmfs1/gscratch/zlab/timdettmers/git/bitsandbytes/csrc/pythonInterface.c with H100 GPU

@yz26cn
Copy link

yz26cn commented Nov 16, 2024

Still getting the issue on H100 with 8-bit setup. 4-bit works as usual but 8-bit does.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working enhancement New feature or request high priority (first issues that will be worked on)
Projects
None yet
Development

No branches or pull requests

10 participants