You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It seems that block sparse attention fails to work after v2.0. It relies on flash_attn_cuda.fwd_block, while flash_attn_cuda has been renamed to flash_attn_2_cuda and has no fwd_block method in it.
Is there any plan to support block sparse attention in v2.0? I can also submit a pull request if I have some direction :)
The text was updated successfully, but these errors were encountered:
Thanks for your great work!
It seems that block sparse attention fails to work after v2.0. It relies on
flash_attn_cuda.fwd_block
, whileflash_attn_cuda
has been renamed toflash_attn_2_cuda
and has nofwd_block
method in it.Is there any plan to support block sparse attention in v2.0? I can also submit a pull request if I have some direction :)
The text was updated successfully, but these errors were encountered: