Skip to content

Commit d491957

Browse files
authored
new modelslim quantization model support v0.7.3 (#743)
### What this PR does / why we need it? In order to support quantization model generated by new modelslim version, we need to add quant_description in `AscendQuantConfig`. cherry-pick this PR from #719 ### Does this PR introduce _any_ user-facing change? no ### How was this patch tested? test locally Signed-off-by: ganyi <[email protected]>
1 parent 595d1a8 commit d491957

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm_ascend/quantization/quant_config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ def get_min_capability(cls) -> int:
7070

7171
@classmethod
7272
def get_config_filenames(cls) -> List[str]:
73-
return []
73+
return ["quant_model_description.json"]
7474

7575
@classmethod
7676
def from_config(cls, config: Dict[str, Any]) -> "AscendQuantConfig":

0 commit comments

Comments
 (0)