Revert "[CI] fix skiped e2e test when upgrade vllm version (#6654)"#7166
Revert "[CI] fix skiped e2e test when upgrade vllm version (#6654)"#7166MengqingCao merged 1 commit intovllm-project:mainfrom
Conversation
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
Summary of ChangesHello, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request reverts a previous change that attempted to address a skipped end-to-end test. Consequently, the Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Changelog
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for GitHub and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Code Review
This PR reverts a previous commit, re-skipping the e2e test test_models_aclgraph_capture_replay_metrics_dp2 to fix CI failures. While this is a valid temporary measure, the skip reason 'fix me' is too vague. I've added a comment suggesting a more descriptive reason to improve tracking and maintainability.
Additionally, the pull request description is empty, which doesn't follow the repository's style guide. Here is a suggested title and summary:
Suggested PR Title:
[Test] Revert "[CI] fix skipped e2e test when upgrade vllm version (#6654)"Suggested PR Summary:
### What this PR does / why we need it?
This PR reverts commit f6db47f1038cf14d6c6f7eb4780be3c839c7656b, which un-skipped the e2e test `test_models_aclgraph_capture_replay_metrics_dp2`. The test is likely failing after a recent vLLM version upgrade, and this revert re-disables it to unblock the CI pipeline.
### Does this PR introduce _any_ user-facing change?
No. This change only affects CI/testing.
### How was this patch tested?
CI is expected to pass with this change, confirming the skipped test was the source of failure.| torch.npu.reset_peak_memory_stats() | ||
|
|
||
|
|
||
| @pytest.mark.skip(reason="fix me") |
There was a problem hiding this comment.
While skipping a test can be a necessary temporary measure, the reason 'fix me' is too generic and makes it difficult to track the problem. Please provide a more descriptive reason, mentioning that this is a revert and why the test is failing. It is also best practice to create and link a tracking issue to ensure the test is re-enabled later.
| @pytest.mark.skip(reason="fix me") | |
| @pytest.mark.skip(reason="Reverting commit f6db47f. Test is failing after vLLM upgrade and needs investigation.") |
…to qwen3next_graph * 'main' of https://github.com/vllm-project/vllm-ascend: (88 commits) [main][bugfix] Fixed the problem of speculative decoding in FULL mode (vllm-project#7148) fixed fia pad logic in graph mode. (vllm-project#7144) [Doc] fix DSV3.1 PD configs (vllm-project#7187) refactor: add a check before layer_sharding logging (vllm-project#7186) [Build] Add support for Ascend950 chip (vllm-project#7151) Revert "[CI] fix skiped e2e test when upgrade vllm version (vllm-project#6654)" (vllm-project#7166) [MODELRUNNERV2]fix penality ops (vllm-project#7013) [Bugfix][LoRA] Fix the issue when enable LoRA + tp + fully_sharded_loras (vllm-project#6650) [KV Pool]get_num_new_matched_tokens return 0 if token length < block_size (vllm-project#7146) [CI] Build Image for v0.16.0rc1 (vllm-project#7155) [CI] Skip `test_mooncake_layerwise_connector.py` in `ut` (vllm-project#7147) [BugFix]Fix recomputed scheduler bug (vllm-project#7137) [Model] Support Minimax-m2.5 on NPU (vllm-project#7105) [P/D]Mooncake Layerwise Connector supports hybrid attention manager with multiple kvcache groups (vllm-project#7022) Add patch_qwen3_5 for triton ops fused_recurrent_gated_delta_rule (vllm-project#7109) [Doc][ReleaseNote] Add release notes for v0.16.0rc1 (vllm-project#7067) [Misc] Download on both hk and guiyang region (vllm-project#7129) [bugdix] The problem that the w4a8 weight fails to be loaded when the EP is not enabled is resolved. (vllm-project#7090) [eagle][cp] fix eagle_cp enable bug2 (vllm-project#7079) [CI]Upgrade niglty multi-node-tests max-parallel to 2 (vllm-project#7035) ...
…ect#6654)" (vllm-project#7166) This reverts commit f6db47f. - vLLM version: v0.16.0 - vLLM main: vllm-project/vllm@4034c3d Signed-off-by: Meihan-chen <jcccx.cmh@gmail.com>
This reverts commit f6db47f.
What this PR does / why we need it?
Does this PR introduce any user-facing change?
How was this patch tested?