-
Notifications
You must be signed in to change notification settings - Fork 91
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Are there any plans to support flash-attn 1? #25
Comments
To anyone struggling with this issue and running on Turing, simply disabling the flash-attn 2 checks in code seems to work just fine! You will then be able to run on flash-attn 1 without errors, and results look similar if not the exact same as hugging-face demo. |
@mateiradu88 How exactly did you do this? I tried |
I deleted the files, but I found an if condition checking for the flash attention version 2, deleting the condition and running the branch anyway worked like a charm, no other modifications as far as I remember. |
Also worked for me, thanks! For anyone else: I just removed the ifelse in line 347 in |
Are there any plans to support flash-attn 1? This is a serious limitation at the moment for anyone still running Turing architecture, as flash-attn 2 does not support Turing.
The text was updated successfully, but these errors were encountered: