Skip to content

Commit

Permalink
Update rocm_flash_attn.py
Browse files Browse the repository at this point in the history
  • Loading branch information
MErkinSag authored Oct 21, 2024
1 parent e107fa4 commit 7deb217
Showing 1 changed file with 0 additions and 3 deletions.
3 changes: 0 additions & 3 deletions vllm/attention/backends/rocm_flash_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -662,9 +662,6 @@ def _sdpa_attention(
def _use_rocm_custom_paged_attention(qtype: torch.dtype, head_size: int,
block_size: int, gqa_ratio: int,
max_seq_len: int) -> bool:
if not _ON_MI250_MI300:
logger.warning(f"Custom Paged Attention is not currently supported on {gpu_arch}.")

# rocm custom page attention not support on navi (gfx1*)
return (_ON_MI250_MI300 and not _ON_NAVI and (qtype == torch.half or qtype == torch.bfloat16)

Check failure on line 666 in vllm/attention/backends/rocm_flash_attn.py

View workflow job for this annotation

GitHub Actions / ruff (3.10)

Ruff (E501)

vllm/attention/backends/rocm_flash_attn.py:666:81: E501 Line too long (97 > 80)
and (head_size == 64 or head_size == 128)
Expand Down

0 comments on commit 7deb217

Please sign in to comment.