Skip to content

Conversation

@hmellor
Copy link
Member

@hmellor hmellor commented Nov 3, 2025

N.B. xformers is also causing --pre to be required at the moment

@hmellor hmellor added the ready ONLY add when PR is ready to merge/full CI is needed label Nov 3, 2025
@mergify mergify bot added the ci/build label Nov 3, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates Flashinfer from version v0.4.1 to v0.5.0. The changes include updating the package versions in requirements/cuda.txt and the Dockerfiles, as well as removing a related workaround that is no longer necessary. The modifications are consistent and correctly implement the version bump. The changes look good.

@youkaichao
Copy link
Member

cc @mgoin @pavanimajety

@simon-mo simon-mo added this to the v0.11.1 milestone Nov 3, 2025
@hmellor
Copy link
Member Author

hmellor commented Nov 3, 2025

Both Blackwell tests passed in last night's nightly. So these appear to be new and legitimate failures.

@lgeiger
Copy link
Contributor

lgeiger commented Nov 4, 2025

Shall we try whether v0.5.1 fixed it?

@hmellor
Copy link
Member Author

hmellor commented Nov 4, 2025

The changelog doesn't look like it fixes any bugs. We have a fix for the unquantised test (relaxing the tolerances), but I'm still waiting for a solution for the quantised moe test.

@hmellor
Copy link
Member Author

hmellor commented Nov 4, 2025

MoE investigation here: flashinfer-ai/flashinfer#2032

@seindum seindum mentioned this pull request Nov 5, 2025
1 task
Signed-off-by: Harry Mellor <[email protected]>
@hmellor
Copy link
Member Author

hmellor commented Nov 7, 2025

flashinfer-ai/flashinfer#2049 contains the MoE fix and was included in 0.5.2 released 1h ago

@hmellor hmellor changed the title Update Flashinfer from v0.4.1 to v0.5.0 Update Flashinfer from v0.4.1 to v0.5.2 Nov 7, 2025
Copy link
Member

@yewentao256 yewentao256 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks for the work!
Could we trigger the full tests (including all optional) for this change?

@hmellor
Copy link
Member Author

hmellor commented Nov 7, 2025

Changing the requirements/dockerfile already does trigger full CI.

There are some optional ones that are only run nightly, but many of those are failing already and so would block this PR from merging.

@hmellor hmellor merged commit 811df41 into vllm-project:main Nov 8, 2025
91 checks passed
@hmellor hmellor deleted the flashinfer-update branch November 8, 2025 00:24
@mgoin mgoin added this to NVIDIA Nov 11, 2025
@mgoin mgoin added the nvidia label Nov 11, 2025
@mgoin mgoin moved this to Done in NVIDIA Nov 11, 2025
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Nov 13, 2025
Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: xuebwang-amd <[email protected]>
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build nvidia ready ONLY add when PR is ready to merge/full CI is needed

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

[Installation]: FlashInfer Dependency issue due to pre-release apache-tvm-ffi

6 participants