Skip to content

Commit caf412e

Browse files
hmellorkhluu
authored andcommitted
Skip models that cannot currently init on Transformers v5 (#28471)
Signed-off-by: Harry Mellor <[email protected]> (cherry picked from commit 51c599f)
1 parent a035b5c commit caf412e

File tree

3 files changed

+6
-16
lines changed

3 files changed

+6
-16
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -868,12 +868,12 @@ steps:
868868
optional: true
869869
commands:
870870
- pip install --upgrade git+https://github.com/huggingface/transformers
871-
- pytest -v -s tests/models/test_initialization.py
871+
- pytest -v -s tests/models/test_initialization.py -k 'not (Gemma3 or ModernBert or Qwen2_5_VL or Qwen2_5vl or Qwen2VL or TransformersMultiModalEmbeddingModel or TransformersMultiModalForSequenceClassification or Ultravox or Phi4Multimodal or LlavaNextVideo or MiniCPMO or Lfm2Moe or PaliGemma or RobertaForSequenceClassification or Ovis2_5 or Fuyu or DeepseekOCR or KimiVL)'
872872
- pytest -v -s tests/models/test_transformers.py
873-
- pytest -v -s tests/models/multimodal/processing/
874-
- pytest -v -s tests/models/multimodal/test_mapping.py
873+
# - pytest -v -s tests/models/multimodal/processing/
874+
- pytest -v -s tests/models/multimodal/test_mapping.py -k 'not (Gemma3 or Qwen2VL or Qwen2_5_VL)'
875875
- python3 examples/offline_inference/basic/chat.py
876-
- python3 examples/offline_inference/vision_language.py --model-type qwen2_5_vl
876+
# - python3 examples/offline_inference/vision_language.py --model-type qwen2_5_vl
877877
# Whisper needs spawn method to avoid deadlock
878878
- VLLM_WORKER_MULTIPROC_METHOD=spawn python3 examples/offline_inference/audio_language.py --model-type whisper
879879

vllm/model_executor/model_loader/weight_utils.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,8 @@ def enable_hf_transfer():
8282

8383
class DisabledTqdm(tqdm):
8484
def __init__(self, *args, **kwargs):
85-
super().__init__(*args, **kwargs, disable=True)
85+
kwargs["disable"] = True
86+
super().__init__(*args, **kwargs)
8687

8788

8889
def get_lock(model_name_or_path: str | Path, cache_dir: str | None = None):

vllm/model_executor/models/whisper.py

Lines changed: 0 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,6 @@
1313
BatchFeature,
1414
WhisperConfig,
1515
WhisperFeatureExtractor,
16-
WhisperProcessor,
1716
)
1817
from transformers.models.whisper.modeling_whisper import sinusoids
1918

@@ -660,16 +659,6 @@ class WhisperProcessingInfo(BaseProcessingInfo):
660659
def get_hf_config(self) -> WhisperConfig:
661660
return self.ctx.get_hf_config(WhisperConfig)
662661

663-
def get_hf_processor(self, **kwargs: object) -> WhisperProcessor:
664-
# HACK: Transformers 4.53.2 has issue with whisper tokenizer to
665-
# initialize processor. We use a monkeypatch to fix it here.
666-
# See: https://github.com/vllm-project/vllm/issues/20224
667-
processor_class = WhisperProcessor
668-
tokenizer_class = ("WhisperTokenizer", "WhisperTokenizerFast")
669-
if processor_class.tokenizer_class != tokenizer_class:
670-
processor_class.tokenizer_class = tokenizer_class
671-
return self.ctx.get_hf_processor(processor_class, **kwargs)
672-
673662
def get_supported_mm_limits(self) -> Mapping[str, int | None]:
674663
return {"audio": 1}
675664

0 commit comments

Comments
 (0)