Skip to content

[MyPy] Enable mypy for vllm/model_executor/layers/#40159

Merged
vllm-bot merged 7 commits intovllm-project:mainfrom
hickeyma:enable-mypy-model-exec-layers
Apr 22, 2026
Merged

[MyPy] Enable mypy for vllm/model_executor/layers/#40159
vllm-bot merged 7 commits intovllm-project:mainfrom
hickeyma:enable-mypy-model-exec-layers

Conversation

@hickeyma
Copy link
Copy Markdown
Contributor

@hickeyma hickeyma commented Apr 17, 2026

Part of #26533

$ pre-commit run -a --hook-stage manual mypy-3.10

Before:

vllm/model_executor/layers/activation.py:671: error: "warning_once" of "_VllmLogger" does not return a value (it only ever returns None)  [func-returns-value]
vllm/model_executor/layers/activation.py:706: error: Need type annotation for "_ACTIVATION_AND_MUL_REGISTRY"  [var-annotated]
vllm/model_executor/layers/activation.py:707: error: Argument 1 to "LazyDict" has incompatible type "dict[str, function]"; expected "dict[str, Callable[[], Never]]"  [arg-type]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:18: error: Item "None" of "DPMetadata | None" has no attribute "get_chunk_sizes_across_dp_rank"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:30: error: Item "None" of "DeviceCommunicatorBase | None" has no attribute "all2all_manager"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:82: error: Argument 1 to "flashinfer_alltoall_dispatch" has incompatible type "All2AllManagerBase | Any | None"; expected "All2AllManagerBase"  [arg-type]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:109: error: Argument 1 to "flashinfer_alltoall_combine" has incompatible type "All2AllManagerBase | Any | None"; expected "All2AllManagerBase"  [arg-type]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:132: error: "All2AllManagerBase" has no attribute "ensure_alltoall_workspace_initialized"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:147: error: "All2AllManagerBase" has no attribute "prepare_workspace_tensor"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:175: error: "All2AllManagerBase" has no attribute "workspace_tensor"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:183: error: "All2AllManagerBase" has no attribute "workspace_tensor"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:199: error: "All2AllManagerBase" has no attribute "workspace_tensor"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:215: error: "All2AllManagerBase" has no attribute "ensure_alltoall_workspace_initialized"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_two_sided.py:221: error: "All2AllManagerBase" has no attribute "workspace_tensor"  [attr-defined]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:14: error: Item "None" of "DPMetadata | None" has no attribute "get_chunk_sizes_across_dp_rank"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:35: error: Item "None" of "DeviceCommunicatorBase | None" has no attribute "all2all_manager"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:36: error: Item "All2AllManagerBase" of "All2AllManagerBase | Any | None" has no attribute "initialize"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:36: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "initialize"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:100: error: Item "All2AllManagerBase" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:100: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:134: error: Item "All2AllManagerBase" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:134: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:136: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "world_size"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:142: error: Item "All2AllManagerBase" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/flashinfer_nvlink_one_sided.py:142: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "moe_alltoall"  [union-attr]
vllm/model_executor/layers/fused_moe/prepare_finalize/naive_dp_ep.py:135: error: Too many values to unpack (3 expected, 4 provided)  [misc]
vllm/model_executor/layers/fused_moe/prepare_finalize/naive_dp_ep.py:138: error: Need more than 3 values to unpack (4 expected)  [misc]
vllm/model_executor/layers/fused_moe/prepare_finalize/naive_dp_ep.py:220: error: Too many values to unpack (2 expected, 3 provided)  [misc]
vllm/model_executor/layers/fused_moe/prepare_finalize/naive_dp_ep.py:223: error: Need more than 2 values to unpack (3 expected)  [misc]
vllm/model_executor/layers/fused_moe/experts/batched_deep_gemm_moe.py:213: error: Item "None" of "DeviceCapability | None" has no attribute "to_int"  [union-attr]
vllm/model_executor/layers/fused_moe/all2all_utils.py:123: error: Item "None" of "DeviceCommunicatorBase | None" has no attribute "all2all_manager"  [union-attr]
vllm/model_executor/layers/fused_moe/all2all_utils.py:123: error: Item "None" of "All2AllManagerBase | Any | None" has no attribute "world_size"  [union-attr]
vllm/model_executor/layers/fused_moe/all2all_utils.py:130: error: Item "None" of "DeviceCommunicatorBase | None" has no attribute "all2all_manager"  [union-attr]
vllm/model_executor/layers/layernorm.py:83: error: Module has no attribute "poly_norm"  [attr-defined]
vllm/model_executor/layers/fused_moe/oracle/unquantized.py:214: error: Argument 1 to "map_unquantized_backend" has incompatible type "str"; expected "Literal['auto', 'triton', 'deep_gemm', 'cutlass', 'flashinfer_trtllm', 'flashinfer_cutlass', 'flashinfer_cutedsl', 'marlin', 'aiter']"  [arg-type]
vllm/model_executor/layers/quantization/fp_quant.py:254: error: Argument "method" to "fusedQuantizeMx" has incompatible type "str"; expected "Literal['quest', 'abs_max']"  [arg-type]
vllm/model_executor/layers/attention/attention.py:217: error: Incompatible types in assignment (expression has type "int | None", variable has type "int")  [assignment]
vllm/model_executor/layers/attention/attention.py:338: error: Incompatible types in assignment (expression has type "AttentionImplBase[Any]", variable has type "AttentionImpl[Any]")  [assignment]
vllm/model_executor/layers/attention/attention.py:700: error: Incompatible types in assignment (expression has type "AttentionMetadata", variable has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]")  [assignment]
vllm/model_executor/layers/fused_moe/runner/default_moe_runner.py:61: error: Too many values to unpack (2 expected, 3 provided)  [misc]
vllm/model_executor/layers/attention/mla_attention.py:392: error: Incompatible types in assignment (expression has type "MLAAttentionImpl[Any]", variable has type "AttentionImpl[Any]")  [assignment]
vllm/model_executor/layers/attention/mla_attention.py:490: error: Incompatible types in assignment (expression has type "AttentionMetadata", variable has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]")  [assignment]
vllm/model_executor/layers/attention/mla_attention.py:497: error: "AttentionImpl[Any]" has no attribute "do_kv_cache_update"; maybe "do_rope_and_kv_cache_update"?  [attr-defined]
vllm/model_executor/layers/attention/mla_attention.py:511: error: Argument 5 to "forward_impl" of "MLAAttention" has incompatible type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]"; expected "MLACommonMetadata[Any]"  [arg-type]
vllm/model_executor/layers/attention/mla_attention.py:615: error: "AttentionImpl[Any]" has no attribute "forward_mha"; maybe "forward"?  [attr-defined]
vllm/model_executor/layers/attention/mla_attention.py:698: error: "AttentionImpl[Any]" has no attribute "forward_mqa"; maybe "forward"?  [attr-defined]
vllm/model_executor/layers/attention/mla_attention.py:1056: error: Name "ops" already defined (by an import)  [no-redef]
vllm/model_executor/layers/attention/mla_attention.py:1058: error: Module has no attribute "flash_attn_varlen_func"  [attr-defined]
vllm/model_executor/layers/attention/mla_attention.py:1058: note: Error code "attr-defined" not covered by "type: ignore" comment
vllm/model_executor/layers/attention/mla_attention.py:1991: error: Incompatible return value type (got "MLACommonMetadata[Any]", expected "M")  [return-value]
vllm/model_executor/layers/fused_moe/oracle/nvfp4.py:211: error: Argument 1 to "map_nvfp4_backend" has incompatible type "str"; expected "Literal['auto', 'triton', 'deep_gemm', 'cutlass', 'flashinfer_trtllm', 'flashinfer_cutlass', 'flashinfer_cutedsl', 'marlin', 'aiter']"  [arg-type]
vllm/model_executor/layers/fused_moe/oracle/fp8.py:276: error: Argument 1 to "map_fp8_backend" has incompatible type "str"; expected "Literal['auto', 'triton', 'deep_gemm', 'cutlass', 'flashinfer_trtllm', 'flashinfer_cutlass', 'flashinfer_cutedsl', 'marlin', 'aiter']"  [arg-type]
vllm/model_executor/layers/attention/encoder_only_attention.py:44: error: Argument "attention_backend_cls" to "subclass_attention_backend" has incompatible type "AttentionBackend"; expected "type[AttentionBackend]"  [arg-type]
vllm/model_executor/layers/attention/encoder_only_attention.py:98: error: Incompatible return value type (got "None", expected "KVCacheSpec")  [return-value]
vllm/model_executor/layers/attention/cross_attention.py:90: error: Item "None" of "Any | None" has no attribute "max"  [union-attr]
vllm/model_executor/layers/attention/cross_attention.py:148: error: "AttentionMetadata" has no attribute "slot_mapping"  [attr-defined]
vllm/model_executor/layers/attention/cross_attention.py:165: error: Argument "attention_backend_cls" to "subclass_attention_backend_with_overrides" has incompatible type "AttentionBackend"; expected "type[AttentionBackend]"  [arg-type]
vllm/model_executor/layers/attention/chunked_local_attention.py:74: error: Argument "attention_backend_cls" to "subclass_attention_backend" has incompatible type "AttentionBackend"; expected "type[AttentionBackend]"  [arg-type]
vllm/model_executor/layers/fused_moe/oracle/mxfp4.py:204: error: Unsupported operand types for <= ("tuple[int, int]" and "None")  [operator]
vllm/model_executor/layers/fused_moe/oracle/mxfp4.py:204: note: Right operand is of type "DeviceCapability | None"
vllm/model_executor/layers/fused_moe/oracle/mxfp4.py:204: error: Unsupported operand types for > ("tuple[int, int]" and "None")  [operator]
vllm/model_executor/layers/fused_moe/oracle/mxfp4.py:204: note: Left operand is of type "DeviceCapability | None"
vllm/model_executor/layers/quantization/quark/quark.py:123: error: Incompatible types in assignment (expression has type "dict[str, Any]", target has type "list[str]")  [assignment]
vllm/model_executor/layers/quantization/quark/quark.py:128: error: Incompatible types in assignment (expression has type "str", target has type "list[str]")  [assignment]
vllm/model_executor/layers/pooler/tokwise/poolers.py:130: error: Item "None" of "PoolerConfig | None" has no attribute "logit_mean"  [union-attr]
vllm/model_executor/layers/pooler/tokwise/poolers.py:131: error: Item "None" of "PoolerConfig | None" has no attribute "logit_sigma"  [union-attr]
vllm/model_executor/layers/pooler/seqwise/poolers.py:121: error: Item "None" of "PoolerConfig | None" has no attribute "logit_mean"  [union-attr]
vllm/model_executor/layers/pooler/seqwise/poolers.py:122: error: Item "None" of "PoolerConfig | None" has no attribute "logit_sigma"  [union-attr]
vllm/model_executor/layers/sparse_attn_indexer.py:29: error: Name "ops" already defined (by an import)  [no-redef]
vllm/model_executor/layers/sparse_attn_indexer.py:87: error: Incompatible types in assignment (expression has type "AttentionMetadata", variable has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]")  [assignment]
vllm/model_executor/layers/sparse_attn_indexer.py:105: error: Argument 5 to "indexer_k_quant_and_cache" has incompatible type "str | None"; expected "str"  [arg-type]
vllm/model_executor/layers/sparse_attn_indexer.py:147: error: Module has no attribute "top_k_per_row_prefill"  [attr-defined]
vllm/model_executor/layers/sparse_attn_indexer.py:224: error: Module has no attribute "top_k_per_row_decode"  [attr-defined]
vllm/model_executor/layers/mamba/abstract.py:47: error: Argument "shapes" to "MambaSpec" has incompatible type "Iterable[tuple[int, ...]]"; expected "tuple[tuple[int, ...], ...]"  [arg-type]
vllm/model_executor/layers/mamba/abstract.py:49: error: Argument "block_size" to "MambaSpec" has incompatible type "int | None"; expected "int"  [arg-type]
vllm/model_executor/layers/mamba/short_conv.py:116: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/mamba/linear_attn.py:399: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/kda.py:126: error: "ModelConfig" has no attribute "linear_attn_config"  [attr-defined]
vllm/model_executor/layers/kda.py:300: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/kda.py:376: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/kda.py:412: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/kda.py:412: error: Unsupported operand type for ~ ("Any | None")  [operator]
vllm/model_executor/layers/kda.py:443: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/mamba_mixer.py:269: error: Incompatible types in assignment (expression has type "AttentionMetadata", variable has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]")  [assignment]
vllm/model_executor/layers/mamba/mamba_mixer.py:395: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer.py:398: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer2.py:575: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/mamba/mamba_mixer2.py:713: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer2.py:752: error: Unsupported operand types for // ("None" and "int")  [operator]
vllm/model_executor/layers/mamba/mamba_mixer2.py:752: note: Left operand is of type "int | None"
vllm/model_executor/layers/mamba/mamba_mixer2.py:757: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/mamba_mixer2.py:776: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/mamba_mixer2.py:785: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/mamba_mixer2.py:793: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/mamba_mixer2.py:814: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer2.py:828: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer2.py:831: error: Item "None" of "Any | None" has no attribute "gather"  [union-attr]
vllm/model_executor/layers/mamba/mamba_mixer2.py:857: error: Item "None" of "Any | None" has no attribute "size"  [union-attr]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:124: error: Item "SupportsHash" of "dict[Any, Any] | SupportsHash" has no attribute "get"  [union-attr]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:616: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:624: error: Value of type "AttentionMetadata" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:627: error: "AttentionMetadata" has no attribute "spec_sequence_masks"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:653: error: "AttentionMetadata" has no attribute "num_prefills"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:654: error: "AttentionMetadata" has no attribute "num_decodes"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:655: error: "AttentionMetadata" has no attribute "has_initial_state"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:656: error: "AttentionMetadata" has no attribute "non_spec_query_start_loc"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:657: error: "AttentionMetadata" has no attribute "non_spec_state_indices_tensor"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:658: error: "AttentionMetadata" has no attribute "num_actual_tokens"  [attr-defined]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:787: error: Incompatible types in assignment (expression has type "dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]]", variable has type "AttentionMetadata")  [assignment]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:861: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:866: error: Item "None" of "Any | None" has no attribute "size"  [union-attr]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:895: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:960: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:974: error: Unsupported operand type for ~ ("Any | None")  [operator]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:1007: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:1068: error: Value of type "Any | None" is not indexable  [index]
vllm/model_executor/layers/mamba/gdn_linear_attn.py:1081: error: Value of type "Any | None" is not indexable  [index]
Found 113 errors in 29 files (checked 1140 source files)

After:

Run mypy for Python 3.10.................................................Passed

Copy link
Copy Markdown

@claude claude Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Claude Code Review

This pull request is from a fork — automated review is disabled. A repository maintainer can comment @claude review to run a one-time review.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request focuses on improving type safety and static analysis across the vllm/model_executor/layers directory, allowing it to be removed from the mypy ignore list. Key changes include adding explicit type hints, introducing assert statements to handle potentially null values, and standardizing the extraction of attn_metadata from the forward context to support speculative decoding. Additionally, the PR addresses platform-specific typing issues for XPU and refactors activation registries for better type compliance. I have no feedback to provide as there are no review comments.

@DarkLight1337 DarkLight1337 requested a review from Isotr0py April 18, 2026 05:39
Copy link
Copy Markdown
Member

@Isotr0py Isotr0py left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Overall look reasonable. Thanks for cleanup!

Comment thread vllm/model_executor/layers/attention/mla_attention.py Outdated
@github-project-automation github-project-automation Bot moved this to Ready in NVIDIA Apr 18, 2026
@Isotr0py Isotr0py added the ready ONLY add when PR is ready to merge/full CI is needed label Apr 18, 2026
@Isotr0py
Copy link
Copy Markdown
Member

Seems tests are failing: https://buildkite.com/vllm/ci/builds/61969#019da16c-3ab1-4d5d-b149-199047fc74fd

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
@hickeyma hickeyma force-pushed the enable-mypy-model-exec-layers branch from fba05eb to d826226 Compare April 20, 2026 08:04
Review comment:

- vllm-project#40159 (comment)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
@Isotr0py Isotr0py enabled auto-merge (squash) April 21, 2026 16:52
@hickeyma
Copy link
Copy Markdown
Contributor Author

Seems tests are failing: https://buildkite.com/vllm/ci/builds/61969#019da16c-3ab1-4d5d-b149-199047fc74fd

@Isotr0py Thank you for the review and feedback. I fixed the issues with the tests. There are other issues with unit tests that seem to be flakiness with the gate at the moment.

@vllm-bot vllm-bot merged commit 3951d3e into vllm-project:main Apr 22, 2026
78 of 80 checks passed
@github-project-automation github-project-automation Bot moved this from Ready to Done in NVIDIA Apr 22, 2026
@hickeyma hickeyma deleted the enable-mypy-model-exec-layers branch April 22, 2026 08:07
Copilot AI pushed a commit to hongbolv/vllm that referenced this pull request Apr 22, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Co-authored-by: hongbolv <33214277+hongbolv@users.noreply.github.com>
baonudesifeizhai pushed a commit to baonudesifeizhai/vllm that referenced this pull request Apr 23, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
yzong-rh pushed a commit to yzong-rh/vllm that referenced this pull request Apr 23, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Signed-off-by: Yifan <yzong@redhat.com>
avinashsingh77 pushed a commit to avinashsingh77/vllm that referenced this pull request Apr 27, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Signed-off-by: Avinash Singh <avinashsingh.rcoem@gmail.com>
Lafunamor pushed a commit to Lafunamor/vllm that referenced this pull request May 1, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Signed-off-by: Adrian <info@zzit.ch>
Copilot AI pushed a commit to hongbolv/vllm that referenced this pull request May 7, 2026
)

Signed-off-by: Martin Hickey <martin.hickey@ie.ibm.com>
Co-authored-by: hongbolv <33214277+hongbolv@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

nvidia ready ONLY add when PR is ready to merge/full CI is needed

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

3 participants