Skip to content

Commit ebed558

Browse files
vllmellmliuzijing2014
authored andcommitted
[BugFix] Revert ROCm Custom Paged Attention Env Flag Check (vllm-project#17022)
Signed-off-by: vllmellm <[email protected]>
1 parent 4192840 commit ebed558

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

vllm/platforms/rocm.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -118,6 +118,7 @@ def use_rocm_custom_paged_attention(qtype: torch.dtype, head_size: int,
118118
and (head_size == 64 or head_size == 128)
119119
and (block_size == 16 or block_size == 32)
120120
and (gqa_ratio >= 1 and gqa_ratio <= 16) and max_seq_len <= 32768
121+
and (envs.VLLM_ROCM_CUSTOM_PAGED_ATTN)
121122
and not (envs.VLLM_ROCM_USE_AITER_PAGED_ATTN
122123
and envs.VLLM_ROCM_USE_AITER))
123124

0 commit comments

Comments
 (0)