Skip to content

[Misc] Upgrade vllm hash to 1210#4906

Closed
Potabk wants to merge 1 commit intovllm-project:mainfrom
Potabk:fix_1210
Closed

[Misc] Upgrade vllm hash to 1210#4906
Potabk wants to merge 1 commit intovllm-project:mainfrom
Potabk:fix_1210

Conversation

@Potabk
Copy link
Copy Markdown
Collaborator

@Potabk Potabk commented Dec 11, 2025

What this PR does / why we need it?

  1. fix [v1] Add PrefixLM support to FlexAttention backend vllm#27938
  2. fix [Model][6/N] Improve all pooling task | Support chunked prefill with ALL pooling vllm#27145
    pooling models now supports chunked prefill and prefix caching,
  3. fix [Model] Move multimodal_cpu_fields definition to field config vllm#30181
    define the CPU fields in the field config where they really belong.
  4. fix [Core][MM] Add mechanism to configure multimodal fields which should stay on CPU vllm#28168
    define the CPU fields in the field config where they really belong.
  5. fix kv_transfer: Rename the shared storage connectors vllm#30201
    some moudle rename
  6. fix [MoE][Refactor] Make select_experts a non-static method vllm#29067
    fusedmoe moudle refactor
  7. fix [MoE][Refactor] Remove most arguments to FusedMoEMethodBase.apply vllm#29066
    fusedmoe moudle refactor
  8. fix [Attention] Make seq_lens_cpu optional in CommonAttentionMetadata to enable true async spec-decode vllm#29624

Does this PR introduce any user-facing change?

How was this patch tested?

Co-authored-by: ZixuanWang 1476209578@qq.com

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request appears to upgrade vLLM compatibility, adapting to upstream changes. The modifications span documentation, tests, and core logic, particularly for pooling, multi-modal features, and speculative decoding. The changes introduce version-specific logic to maintain backward compatibility with vLLM v0.12.0. Overall, the changes are well-structured. I have one comment regarding a potentially misleading comment in the resource calculation logic, which could impact future maintainability.

Comment thread vllm_ascend/utils.py
@github-actions
Copy link
Copy Markdown
Contributor

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@Potabk Potabk added ready read for review ready-for-test start test by label for PR labels Dec 11, 2025
@github-actions
Copy link
Copy Markdown
Contributor

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@Potabk
Copy link
Copy Markdown
Collaborator Author

Potabk commented Dec 11, 2025

unit tests passed locally:
image

@Potabk
Copy link
Copy Markdown
Collaborator Author

Potabk commented Dec 11, 2025

@Potabk
Copy link
Copy Markdown
Collaborator Author

Potabk commented Dec 11, 2025

@github-actions
Copy link
Copy Markdown
Contributor

This pull request has conflicts, please resolve those before we can evaluate the pull request.

Signed-off-by: wangli <wangli858794774@gmail.com>
@github-actions
Copy link
Copy Markdown
Contributor

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@Potabk Potabk closed this Dec 15, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation merge-conflicts module:core module:ops module:tests ready read for review ready-for-test start test by label for PR

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant