Skip to content

[Test][e2e][LoRA] Add more e2e tests to cover scenarios of LoRA#4075

Merged
paulyu12 merged 2 commits intovllm-project:mainfrom
paulyu12:lora_llama_testcase
Jan 13, 2026
Merged

[Test][e2e][LoRA] Add more e2e tests to cover scenarios of LoRA#4075
paulyu12 merged 2 commits intovllm-project:mainfrom
paulyu12:lora_llama_testcase

Conversation

@paulyu12
Copy link
Copy Markdown
Collaborator

@paulyu12 paulyu12 commented Nov 10, 2025

What this PR does / why we need it?

This PR depends on PR #4046. And only if the latter merged, it will work.

This PR aims to solve the issue #3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the senarios that the LoRA weights are added to q_proj, v_proj, k_proj, o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

Does this PR introduce any user-facing change?

No.

How was this patch tested?

pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py

@github-actions
Copy link
Copy Markdown
Contributor

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds end-to-end tests for LoRA scenarios. The changes are mostly good, but I've found a high-severity issue in tests/e2e/singlecard/test_qwen3_multi_loras.py related to test isolation. The test uses mutable global state, which can lead to flaky and hard-to-maintain tests. I've provided a refactoring suggestion to address this by encapsulating the state within the test function. This will make the test self-contained and robust.

Comment thread tests/e2e/singlecard/test_qwen3_multi_loras.py Outdated
@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR labels Nov 11, 2025
@github-actions
Copy link
Copy Markdown
Contributor

github-actions bot commented Dec 1, 2025

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@paulyu12 paulyu12 force-pushed the lora_llama_testcase branch from a0b213b to b2a1cb8 Compare December 22, 2025 09:04
@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR and removed ready read for review ready-for-test start test by label for PR labels Dec 22, 2025
@github-actions
Copy link
Copy Markdown
Contributor

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@wangxiyuan
Copy link
Copy Markdown
Collaborator

@paulyu12 any update?

@paulyu12
Copy link
Copy Markdown
Collaborator Author

paulyu12 commented Jan 5, 2026

@paulyu12 any update?

This CI issue was introduced by #4168 .I tried according to your instruction (Don't make LoRA scenario go into this if condition statement), but it never works.

I am still working on this.

@wangxiyuan
Copy link
Copy Markdown
Collaborator

@wxsIcey please take a look

@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR and removed ready read for review ready-for-test start test by label for PR labels Jan 8, 2026
@paulyu12 paulyu12 removed ready read for review ready-for-test start test by label for PR labels Jan 9, 2026
@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR labels Jan 9, 2026
@github-actions
Copy link
Copy Markdown
Contributor

github-actions bot commented Jan 9, 2026

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR and removed ready read for review ready-for-test start test by label for PR labels Jan 9, 2026
Signed-off-by: paulyu12 <507435917@qq.com>
@paulyu12 paulyu12 force-pushed the lora_llama_testcase branch from 142c570 to 825e2d2 Compare January 12, 2026 03:10
Signed-off-by: paulyu12 <507435917@qq.com>
@paulyu12 paulyu12 added ready read for review ready-for-test start test by label for PR and removed ready read for review ready-for-test start test by label for PR labels Jan 12, 2026
@paulyu12 paulyu12 merged commit 5b95c6b into vllm-project:main Jan 13, 2026
61 of 62 checks passed
845473182 pushed a commit to 845473182/vllm-ascend that referenced this pull request Jan 14, 2026
…to eplb_refactor

* 'main' of https://github.com/vllm-project/vllm-ascend:
  [CI] Fix lint CI (vllm-project#5880)
  [Feature] implement eagle spec decoding for model runner v2 (vllm-project#5840)
  [Quantization] Support compressed tensors moe w8a8 int8 dynamic weight (vllm-project#5718)
  [EPLB][Bugfix] Get expert map from layers (vllm-project#5817)
  [Bugfix] Fixed an accuracy problem of sp with eagle3 (vllm-project#5816)
  [P/D] bugfix for p node force free requset (vllm-project#5431)
  [Lint]Style: Convert `example` to `ruff format` (vllm-project#5863)
  [Main2Main] Upgrade vllm commit to 0109 (vllm-project#5752)
  [Bugfix][P/D] fix layerwise connector for decoder tp size > num kv heads (vllm-project#5846)
  [Test][e2e][LoRA] Add more e2e tests to cover scenarios of LoRA (vllm-project#4075)
  [CustomOp][Perf] Merge Q/K split to simplify AscendApplyRotaryEmb for better performance (vllm-project#5799)
  [Lint]Style: Convert `root`, `benchmarks`, `tools` and `docs` to `ruff format` (vllm-project#5843)
  enable ep32 for dispatch_ffn_combine (vllm-project#5787)
aipaes pushed a commit to aipaes/vllm-ascend that referenced this pull request Jan 15, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
starmountain1997 pushed a commit to starmountain1997/vllm-ascend that referenced this pull request Jan 31, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
starmountain1997 pushed a commit to starmountain1997/vllm-ascend that referenced this pull request Jan 31, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Feb 28, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
maoxx241 pushed a commit to maoxx241/vllm-ascend that referenced this pull request Mar 2, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Mar 4, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
LCAIZJ pushed a commit to LCAIZJ/vllm-ascend that referenced this pull request Mar 7, 2026
…-project#4075)

### What this PR does / why we need it?

This PR depends on PR
vllm-project#4046. And only if the
latter merged, it will work.

This PR aims to solve the issue
vllm-project#3240.

The new-added Llama-2-7b-hf and Qwen3-0.6B testcases will cover the
senarios that the LoRA weights are added to q_proj, v_proj, k_proj,
o_proj, gate_proj, up_proj, down_proj, embed_tokens and lm_head modules.

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
pytest -sv tests/e2e/singlecard/test_llama2_lora.py
pytest -sv tests/e2e/singlecard/test_qwen3_multi_loras.py


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: paulyu12 <507435917@qq.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

module:tests ready read for review ready-for-test start test by label for PR

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants