-
Hello, I was wondering, is it possible to run flash (or better) attention on Turing GPUs (like T4) with your framework? I would much rather use TPUs obviously, but Kaggle competitions allow only GPUs for submission and we really need large context. Pytorch currently doesn't have support for these GPUs. |
Beta Was this translation helpful? Give feedback.
Answered by
erfanzar
May 6, 2024
Replies: 2 comments
-
Hello, I'm working on that right now but, yes soon I assume it will be possible |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
defdet
-
Looking forward to it, that would help a lot! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hello, I'm working on that right now but, yes soon I assume it will be possible