Skip to content

Commit

Permalink
Fix bug in mem_eff kernel with attention mask and MQA (#952)
Browse files Browse the repository at this point in the history
  • Loading branch information
drisspg authored Feb 23, 2024
1 parent 292be24 commit 4f7ffaa
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions xformers/csrc/attention/cuda/fmha/kernel_forward.h
Original file line number Diff line number Diff line change
Expand Up @@ -327,6 +327,7 @@ struct AttentionKernel {
return false;
}
q_strideM = q_strideH;
bias_strideM = bias_strideH;
num_queries = num_heads;
num_heads = 1; // unused but here for intent
// remove causal since n_query = 1
Expand Down

0 comments on commit 4f7ffaa

Please sign in to comment.