Skip to content

[Bugfix] Fix Plamo3 rope handling#29092

Merged
DarkLight1337 merged 4 commits intovllm-project:mainfrom
DarkLight1337:fix-plamo3
Nov 21, 2025
Merged

[Bugfix] Fix Plamo3 rope handling#29092
DarkLight1337 merged 4 commits intovllm-project:mainfrom
DarkLight1337:fix-plamo3

Conversation

@DarkLight1337
Copy link
Copy Markdown
Member

@DarkLight1337 DarkLight1337 commented Nov 20, 2025

Purpose

Fix a failure in model initialization tests caused by an incompatibility between #28834 and #28542 .

https://buildkite.com/vllm/ci/builds/39885#019aa0ed-01f7-4df6-a93b-2d95f144776c

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@DarkLight1337 DarkLight1337 added the ready ONLY add when PR is ready to merge/full CI is needed label Nov 20, 2025
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request aims to fix a bug in the Plamo3 model's RoPE handling. While the change fixes a TypeError from an updated function signature, it introduces a critical correctness bug by removing layer-specific logic for rope_theta. My review includes a comment with a code suggestion to fix this new bug by correctly preserving the dynamic rope_theta selection.

Copy link
Copy Markdown

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@DarkLight1337
Copy link
Copy Markdown
Member Author

DarkLight1337 commented Nov 20, 2025

@Alnusjaponica can you check if this works? I copied the RoPE initialization code from Gemma3 model.

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@hmellor
Copy link
Copy Markdown
Member

hmellor commented Nov 20, 2025

Is Plamo2's config implemented in vLLM?

If yes it can be updated to create a nested paramaters_config (see Gemma 3 on Transformers main for reference) and we don't need to do the init time check in the model code, it would always be nested.

@DarkLight1337
Copy link
Copy Markdown
Member Author

No, it's only in vLLM's modeling file for type annotation purposes

Copy link
Copy Markdown
Member

@hmellor hmellor left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Overall LGTM, just one comment on something I missed in Gemma 3 which someone has a fix for.


Side notes about v5 withthis model:

  • It look like it's a custom model, which means that unless the model authors update their repo, it will never support v5 style RoPE.
  • However, including the v5 block could be a good way to convince them to update it, because there would be no downtime where their model was unsupported.

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@DarkLight1337 DarkLight1337 merged commit 0e741c1 into vllm-project:main Nov 21, 2025
50 checks passed
@DarkLight1337 DarkLight1337 deleted the fix-plamo3 branch November 21, 2025 03:38
ywang96 pushed a commit to ywang96/vllm that referenced this pull request Nov 23, 2025
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
RunkaiTao pushed a commit to RunkaiTao/vllm that referenced this pull request Nov 24, 2025
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Signed-off-by: Runkai Tao <rt572@physics.rutgers.edu>
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
kitaekatt pushed a commit to kitaekatt/vllm that referenced this pull request Dec 1, 2025
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants