Skip to content

Commit e99d368

Browse files
vllmellmamitm02
authored andcommitted
[ROCm] Remove unnecessary assertion of max_model_len in ROCM_AITER_MLA attention backend. (vllm-project#18938)
Signed-off-by: vllmellm <[email protected]> Signed-off-by: amit <[email protected]>
1 parent 9ed94b4 commit e99d368

File tree

2 files changed

+0
-5
lines changed

2 files changed

+0
-5
lines changed

vllm/attention/backends/rocm_aiter_mla.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -132,8 +132,6 @@ class AiterMLAMetadataBuilder(MLACommonMetadataBuilder[AiterMLAMetadata]):
132132

133133
def __init__(self, input_builder: "ModelInputForGPUBuilder"):
134134
super().__init__(input_builder)
135-
assert self.runner.model_config.max_model_len == 32768,\
136-
"AITER MLA requires max model len to be set to 32768"
137135
assert self.block_size == 1, "AITER MLA requires only block size 1."
138136

139137
def prepare(self):

vllm/v1/attention/backends/mla/rocm_aiter_mla.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -66,9 +66,6 @@ class AiterMLAMetadataBuilder(MLACommonMetadataBuilder[AiterMLAMetadata]):
6666
def __init__(self, runner, kv_cache_spec: AttentionSpec,
6767
block_table: BlockTable):
6868
super().__init__(runner, kv_cache_spec, block_table)
69-
max_model_len = self.runner.model_config.max_model_len
70-
assert max_model_len == 32768,\
71-
"AITER MLA requires max_model_len=32768"
7269
assert self.kv_cache_spec.block_size == 1, "AITER MLA" \
7370
"only supports block size 1."
7471

0 commit comments

Comments
 (0)