Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 2 additions & 3 deletions vllm/model_executor/layers/quantization/fp8.py
Original file line number Diff line number Diff line change
Expand Up @@ -62,10 +62,9 @@ def __init__(
weight_block_size: Optional[list[int]] = None,
) -> None:
super().__init__()

self.is_checkpoint_fp8_serialized = is_checkpoint_fp8_serialized
if is_checkpoint_fp8_serialized:
logger.warning("Detected fp8 checkpoint. Please note that the "
"format is experimental and subject to change.")

if activation_scheme not in ACTIVATION_SCHEMES:
raise ValueError(
f"Unsupported activation scheme {activation_scheme}")
Expand Down
5 changes: 1 addition & 4 deletions vllm/platforms/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -217,11 +217,8 @@ def resolve_current_platform_cls_qualname() -> str:
platform_cls_qualname = func()
if platform_cls_qualname is not None:
activated_plugins.append(name)
logger.info("Platform plugin %s loaded.", name)
logger.warning(
"Platform plugin %s function's return value is None", name)
except Exception:
logger.exception("Failed to load platform plugin %s", name)
pass

activated_builtin_plugins = list(
set(activated_plugins) & set(builtin_platform_plugins.keys()))
Expand Down
Loading