We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4b817cd commit cb1b16cCopy full SHA for cb1b16c
vllm/v1/attention/backends/mla/rocm_aiter_mla.py
@@ -18,6 +18,7 @@
18
MLACommonImpl,
19
MLACommonMetadata,
20
MLACommonMetadataBuilder,
21
+ QueryLenSupport,
22
)
23
from vllm.v1.attention.backends.utils import AttentionCGSupport
24
from vllm.v1.kv_cache_interface import AttentionSpec
@@ -82,6 +83,7 @@ class AiterMLAMetadataBuilder(MLACommonMetadataBuilder[AiterMLAMetadata]):
82
83
cudagraph_support: ClassVar[AttentionCGSupport] = (
84
AttentionCGSupport.UNIFORM_SINGLE_TOKEN_DECODE
85
86
+ query_len_support: ClassVar[QueryLenSupport] = QueryLenSupport.VARLEN
87
88
def __init__(
89
self,
0 commit comments