-
-
Notifications
You must be signed in to change notification settings - Fork 11.8k
[V0 Deprecation] Remove unused context_len and seq_len from M-RoPE
#28395
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[V0 Deprecation] Remove unused context_len and seq_len from M-RoPE
#28395
Conversation
Signed-off-by: DarkLight1337 <[email protected]>
cb698b7 to
cf9ba38
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request removes the unused context_len and seq_len arguments from get_mrope_input_positions across multiple model files, which is a good cleanup as part of the V0 deprecation. The changes are mostly consistent and correct.
However, I've found two critical issues that need to be addressed:
- The
SupportsMRoPEprotocol invllm/model_executor/models/interfaces.pystill definescontext_lenandseq_lenin itsget_mrope_input_positionssignature. This protocol should be updated to match the new function signatures in the implementing classes to avoid signature mismatches. - In
vllm/model_executor/models/qwen3_omni_moe_thinker.py, only the function signature was updated, but the usage ofcontext_lenandseq_lenwithin the function body appears to have been missed. This will likely cause aNameErrorat runtime.
Please address these issues to ensure the correctness and consistency of the codebase.
Signed-off-by: DarkLight1337 <[email protected]>
|
/gemini review |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request aims to remove the unused context_len and seq_len arguments from get_mrope_input_positions functions across multiple models, which is a good cleanup. The changes are mostly correct and consistent. However, I've found a critical issue in qwen3_omni_moe_thinker.py where removing seq_len from the function signature without updating its usage in the function body will lead to a runtime error. Please address this to ensure the code remains functional.
vllm-project#28395) Signed-off-by: DarkLight1337 <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
vllm-project#28395) Signed-off-by: DarkLight1337 <[email protected]>
Purpose
From my understanding, these arguments are only passed in V0, as the slicing is now handled by model runner inside V1, so we can remove this from modeling code.
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.