Skip to content

Fix RoPE related failures in Transformers nightly tests#29333

Merged
hmellor merged 5 commits intovllm-project:mainfrom
hmellor:fix-transformers-nightly
Nov 25, 2025
Merged

Fix RoPE related failures in Transformers nightly tests#29333
hmellor merged 5 commits intovllm-project:mainfrom
hmellor:fix-transformers-nightly

Conversation

@hmellor
Copy link
Member

@hmellor hmellor commented Nov 24, 2025

Should fix the RoPE related failures in the Transformers nightly tests by:

  • Handling rope_theta when loading v4 custom models in v5
  • Patching original_max_position_embeddings for v4 (all models) and v5 (v4 custom models)
  • Only add config.rope_parameters if it was present in the config
  • Don't assume that config.rope_parameters is always available in model code

Test plan:

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@hmellor hmellor added the ready ONLY add when PR is ready to merge/full CI is needed label Nov 24, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request aims to fix an issue with RoPE parameter handling for nightly builds of the transformers library. The change correctly identifies the need to move rope_theta into the rope_parameters dictionary for newer transformers versions. However, the current implementation has a flaw in how it handles interleaved models, which could lead to runtime errors. I've provided a critical review comment with a code suggestion to address this potential bug and ensure the fix is robust for different model architectures.

Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@hmellor hmellor changed the title Fix Transformers nightly Fix RoPE related failures in Transformers nightly tests Nov 25, 2025
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@mergify mergify bot added the llama Related to Llama models label Nov 25, 2025
@hmellor hmellor enabled auto-merge (squash) November 25, 2025 15:59
@hmellor hmellor merged commit 0353d2e into vllm-project:main Nov 25, 2025
52 of 53 checks passed
@hmellor hmellor deleted the fix-transformers-nightly branch November 25, 2025 16:23
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
…#29333)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
kitaekatt pushed a commit to kitaekatt/vllm that referenced this pull request Dec 1, 2025
…#29333)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
…#29333)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

llama Related to Llama models ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants