update flashinfer to v0.2.9rc1#21485
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
There was a problem hiding this comment.
Code Review
This pull request updates the FlashInfer dependency to version v0.2.9rc1. The code changes correctly adapt the calls to trtllm_batch_decode_with_kv_cache to match the updated API in the new version. The changes are consistent across the affected files and appear to be correct.
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com>
aeec6ab to
6289473
Compare
|
This pull request has merge conflicts that must be resolved before it can be |
|
#21408 is already merged. No more changes needed. |
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com> Signed-off-by: x22x22 <wadeking@qq.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com> Signed-off-by: Jinzhen Lin <linjinzhen@hotmail.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com> Signed-off-by: Paul Pak <paulpak58@gmail.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com> Signed-off-by: Diego-Castan <diego.castan@ibm.com>
Signed-off-by: Weiliang Liu <weiliangl@nvidia.com>
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.Purpose
updata flashinfer and modify it's trtllm-gen call to use latest API.
Test Plan
test llama4 with lm_eval
Test Result
(Optional) Documentation Update