Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions tests/entrypoints/offline_mode/test_offline_mode.py
Original file line number Diff line number Diff line change
Expand Up @@ -112,6 +112,7 @@ def _re_import_modules():
aliased_module_patterns = [
r".+\.tokenization_utils$",
r".+\.tokenization_utils_fast$",
r".+\.image_processing_utils_fast$",
r".+\.models\..+\.image_processing_.+_fast$",
]

Expand Down
8 changes: 8 additions & 0 deletions vllm/config/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -586,6 +586,14 @@ def __post_init__(
config_format=self.config_format,
)

# Some checkpoints set sliding_window to 0 to indicate that sliding window is
# disabled, but vLLM uses None for that. Convert 0 to None to avoid errors.
# Set before get_and_verify_max_len to ensure that max_model_len does not get
# capped to 0.
if self.get_sliding_window() == 0:
self.disable_sliding_window = True
self.hf_text_config.sliding_window = None

self.original_max_model_len = self.max_model_len
self.max_model_len = self.get_and_verify_max_len(self.max_model_len)

Expand Down
3 changes: 1 addition & 2 deletions vllm/model_executor/models/olmo2.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@

import torch
from torch import nn
from transformers import Olmo2Config
from transformers import Olmo2Config, Olmo3Config

from vllm.compilation.decorators import support_torch_compile
from vllm.config import VllmConfig
Expand Down Expand Up @@ -63,7 +63,6 @@
maybe_prefix,
)
from vllm.sequence import IntermediateTensors
from vllm.transformers_utils.configs.olmo3 import Olmo3Config


class Olmo2Attention(nn.Module):
Expand Down
1 change: 0 additions & 1 deletion vllm/transformers_utils/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -105,7 +105,6 @@ def __getitem__(self, key):
eagle="EAGLEConfig",
speculators="SpeculatorsConfig",
nemotron="NemotronConfig",
olmo3="Olmo3Config",
olmo_hybrid="OlmoHybridConfig",
ovis="OvisConfig",
ultravox="UltravoxConfig",
Expand Down
2 changes: 0 additions & 2 deletions vllm/transformers_utils/configs/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,6 @@
"KimiK25Config": "vllm.transformers_utils.configs.kimi_k25",
"NemotronConfig": "vllm.transformers_utils.configs.nemotron",
"NemotronHConfig": "vllm.transformers_utils.configs.nemotron_h",
"Olmo3Config": "vllm.transformers_utils.configs.olmo3",
"OlmoHybridConfig": "vllm.transformers_utils.configs.olmo_hybrid",
"OvisConfig": "vllm.transformers_utils.configs.ovis",
"PixelShuffleSiglip2VisionConfig": "vllm.transformers_utils.configs.isaac",
Expand Down Expand Up @@ -106,7 +105,6 @@
"KimiK25Config",
"NemotronConfig",
"NemotronHConfig",
"Olmo3Config",
"OlmoHybridConfig",
"OvisConfig",
"PixelShuffleSiglip2VisionConfig",
Expand Down
3 changes: 3 additions & 0 deletions vllm/transformers_utils/configs/deepseek_vl2.py
Original file line number Diff line number Diff line change
Expand Up @@ -114,6 +114,9 @@ def __init__(
self.projector_config = MlpProjectorConfig(**projector_config)

language_config = kwargs.get("language_config", {})
# remove kv_lora_rank if not specified, passing None is prohibited
if language_config.get("kv_lora_rank") is None:
language_config.pop("kv_lora_rank", None)
self.text_config = DeepseekV2Config(**language_config)
Comment on lines +117 to 120
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

Modifying the language_config dictionary in-place can lead to unexpected side effects for the caller if they reuse the kwargs dictionary. It's safer to work with a copy of the dictionary to avoid such issues.

Suggested change
# remove kv_lora_rank if not specified, passing None is prohibited
if language_config.get("kv_lora_rank") is None:
language_config.pop("kv_lora_rank", None)
self.text_config = DeepseekV2Config(**language_config)
# remove kv_lora_rank if not specified, passing None is prohibited
language_config_copy = language_config.copy()
if language_config_copy.get("kv_lora_rank") is None:
language_config_copy.pop("kv_lora_rank", None)
self.text_config = DeepseekV2Config(**language_config_copy)


self.tile_tag = tile_tag
Expand Down
83 changes: 0 additions & 83 deletions vllm/transformers_utils/configs/olmo3.py

This file was deleted.

Loading