[Doc] Improve MM models LoRA notes#31979
Conversation
|
Documentation preview: https://vllm--31979.org.readthedocs.build/en/31979/ |
There was a problem hiding this comment.
Code Review
This pull request significantly improves the documentation for multimodal (MM) models and their LoRA support. It updates the compatibility tables for several models, including DotsOCR, Grok2, IQuestCoderV1, IQuestLoopCoderV1, LFM2-VL, Pixtral, Qwen3-VL, and Qwen3-VL-Reranker, reflecting expanded LoRA and parallelism capabilities. Notably, the detailed Python code for merging LoRA weights in multimodal encoders has been removed, replaced by a more concise note indicating experimental direct support for LoRA in tower and connector modules. This suggests an enhancement in vLLM's native capabilities, simplifying the user experience. The addition of specific --hf_overrides instructions for Grok2 and Qwen3-VL-Reranker further aids users in configuring these models correctly. Overall, these changes enhance clarity and provide up-to-date information on multimodal model integration and LoRA usage within vLLM.
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Purpose
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.