@@ -314,6 +314,8 @@ Specified using `--task generate`.
314314| ` DeepseekForCausalLM ` | DeepSeek | ` deepseek-ai/deepseek-llm-67b-base ` , ` deepseek-ai/deepseek-llm-7b-chat ` etc. | | ✅︎ |
315315| ` DeepseekV2ForCausalLM ` | DeepSeek-V2 | ` deepseek-ai/DeepSeek-V2 ` , ` deepseek-ai/DeepSeek-V2-Chat ` etc. | | ✅︎ |
316316| ` DeepseekV3ForCausalLM ` | DeepSeek-V3 | ` deepseek-ai/DeepSeek-V3-Base ` , ` deepseek-ai/DeepSeek-V3 ` etc. | | ✅︎ |
317+ | ` Ernie4_5ForCausalLM ` | Ernie4.5 | ` baidu/ERNIE-4.5-0.3B-PT ` ,etc. | | ✅︎ | ✅︎ |
318+ | ` Ernie4_5_MoeForCausalLM ` | Ernie4.5MoE | ` baidu/ERNIE-4.5-21B-A3B-PT ` , ` baidu/ERNIE-4.5-300B-A47B-PT ` , etc. | | ✅︎ | ✅︎ |
317319| ` ExaoneForCausalLM ` | EXAONE-3 | ` LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct ` , etc. | ✅︎ | ✅︎ |
318320| ` FalconForCausalLM ` | Falcon | ` tiiuae/falcon-7b ` , ` tiiuae/falcon-40b ` , ` tiiuae/falcon-rw-7b ` , etc. | | ✅︎ |
319321| ` FalconMambaForCausalLM ` | FalconMamba | ` tiiuae/falcon-mamba-7b ` , ` tiiuae/falcon-mamba-7b-instruct ` , etc. | | ✅︎ |
@@ -371,7 +373,6 @@ Specified using `--task generate`.
371373| ` XverseForCausalLM ` | XVERSE | ` xverse/XVERSE-7B-Chat ` , ` xverse/XVERSE-13B-Chat ` , ` xverse/XVERSE-65B-Chat ` , etc. | ✅︎ | ✅︎ |
372374| ` MiniMaxText01ForCausalLM ` | MiniMax-Text | ` MiniMaxAI/MiniMax-Text-01 ` , etc. | | |
373375| ` Zamba2ForCausalLM ` | Zamba2 | ` Zyphra/Zamba2-7B-instruct ` , ` Zyphra/Zamba2-2.7B-instruct ` , ` Zyphra/Zamba2-1.2B-instruct ` , etc. | | |
374-
375376!!! note
376377 Currently, the ROCm version of vLLM supports Mistral and Mixtral only for context lengths up to 4096.
377378
@@ -556,10 +557,10 @@ Specified using `--task generate`.
556557| ` SmolVLMForConditionalGeneration ` | SmolVLM2 | T + I | ` SmolVLM2-2.2B-Instruct ` | ✅︎ | | ✅︎ |
557558| ` TarsierForConditionalGeneration ` | Tarsier | T + I<sup >E+</sup > | ` omni-search/Tarsier-7b ` ,` omni-search/Tarsier-34b ` | | ✅︎ | ✅︎ |
558559
559- <sup >^</sup > You need to set the architecture name via ` --hf-overrides ` to match the one in vLLM.
560-   ;  ;  ;  ; • For example, to use DeepSeek-VL2 series models:
561-   ;  ;  ;  ;  ;  ; ` --hf-overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}' `
562- <sup >E</sup > Pre-computed embeddings can be inputted for this modality.
560+ <sup >^</sup > You need to set the architecture name via ` --hf-overrides ` to match the one in vLLM.
561+   ;  ;  ;  ; • For example, to use DeepSeek-VL2 series models:
562+   ;  ;  ;  ;  ;  ; ` --hf-overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}' `
563+ <sup >E</sup > Pre-computed embeddings can be inputted for this modality.
563564<sup >+</sup > Multiple items can be inputted per text prompt for this modality.
564565
565566!!! warning
0 commit comments