fix amd triton mla path (#17871)

This commit is contained in:
Shiyan Deng
2025-05-11 00:53:31 -07:00
committed by GitHub
parent 9112155283
commit eea22a56ab

View File

@ -1063,7 +1063,7 @@ class MLACommonImpl(MLAAttentionImpl[T], Generic[T]):
softmax_scale,
None, # bias
)
if is_vllm_fa:
elif is_vllm_fa:
attn_out = self.flash_attn_varlen_func(
q=q,
k=k,