Skip to content

[Test][Accuracy] Add accuracy evaluation config for InternVL3_5-8B#3964

Merged
MengqingCao merged 1 commit intovllm-project:mainfrom
gcanlin:acc-internvl3_5
Nov 12, 2025
Merged

[Test][Accuracy] Add accuracy evaluation config for InternVL3_5-8B#3964
MengqingCao merged 1 commit intovllm-project:mainfrom
gcanlin:acc-internvl3_5

Conversation

@gcanlin
Copy link
Copy Markdown
Collaborator

@gcanlin gcanlin commented Nov 4, 2025

What this PR does / why we need it?

To continuously monitor the accuracy of the InternVL3_5-8B model, this PR adds the corresponding configuration file to the CI. We need to add the -hf suffix to avoid incompatibility with the lm-eval preprocessor.

Does this PR introduce any user-facing change?

None.

How was this patch tested?

pytest -sv ./tests/e2e/models/test_lm_eval_correctness.py --config ./tests/e2e/models/configs/InternVL3_5-8B.yaml

@github-actions
Copy link
Copy Markdown
Contributor

github-actions bot commented Nov 4, 2025

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds a new test configuration file to evaluate the accuracy of the InternVL3_5-8B model. While the configuration is straightforward and appears to serve its purpose, it introduces a critical security vulnerability by enabling trust_remote_code. This setting can lead to arbitrary code execution in the CI environment. My review includes a critical comment detailing this risk and suggesting potential mitigations.

Comment thread tests/e2e/models/configs/InternVL3_5-8B.yaml
@github-actions
Copy link
Copy Markdown
Contributor

github-actions bot commented Nov 4, 2025

This pull request has conflicts, please resolve those before we can evaluate the pull request.

Signed-off-by: gcanlin <canlinguosdu@gmail.com>
@MengqingCao MengqingCao added ready read for review ready-for-test start test by label for PR accuracy-test enable all accuracy test for PR and removed ready read for review labels Nov 8, 2025
Copy link
Copy Markdown
Collaborator

@MengqingCao MengqingCao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@gcanlin
Copy link
Copy Markdown
Collaborator Author

gcanlin commented Nov 12, 2025

@MengqingCao CI has passed. Could you please merge this PR?

@MengqingCao MengqingCao merged commit 1c677c3 into vllm-project:main Nov 12, 2025
60 of 61 checks passed
luolun pushed a commit to luolun/vllm-ascend that referenced this pull request Nov 19, 2025
…llm-project#3964)

### What this PR does / why we need it?

To continuously monitor the accuracy of the InternVL3_5-8B model, this
PR adds the corresponding configuration file to the CI. We need to add
the `-hf` suffix to avoid incompatibility with the `lm-eval`
preprocessor.

### How was this patch tested?

`pytest -sv ./tests/e2e/models/test_lm_eval_correctness.py --config
./tests/e2e/models/configs/InternVL3_5-8B.yaml`


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

Signed-off-by: gcanlin <canlinguosdu@gmail.com>
Signed-off-by: luolun <luolun1995@cmbchina.com>
hwhaokun pushed a commit to hwhaokun/vllm-ascend that referenced this pull request Nov 19, 2025
…llm-project#3964)

### What this PR does / why we need it?

To continuously monitor the accuracy of the InternVL3_5-8B model, this
PR adds the corresponding configuration file to the CI. We need to add
the `-hf` suffix to avoid incompatibility with the `lm-eval`
preprocessor.

### How was this patch tested?

`pytest -sv ./tests/e2e/models/test_lm_eval_correctness.py --config
./tests/e2e/models/configs/InternVL3_5-8B.yaml`

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

Signed-off-by: gcanlin <canlinguosdu@gmail.com>
Signed-off-by: hwhaokun <haokun0405@163.com>
NSDie pushed a commit to NSDie/vllm-ascend that referenced this pull request Nov 24, 2025
…llm-project#3964)

### What this PR does / why we need it?

To continuously monitor the accuracy of the InternVL3_5-8B model, this
PR adds the corresponding configuration file to the CI. We need to add
the `-hf` suffix to avoid incompatibility with the `lm-eval`
preprocessor.

### How was this patch tested?

`pytest -sv ./tests/e2e/models/test_lm_eval_correctness.py --config
./tests/e2e/models/configs/InternVL3_5-8B.yaml`

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

Signed-off-by: gcanlin <canlinguosdu@gmail.com>
Signed-off-by: nsdie <yeyifan@huawei.com>
Clorist33 pushed a commit to Clorist33/vllm-ascend that referenced this pull request Dec 10, 2025
…llm-project#3964)

### What this PR does / why we need it?

To continuously monitor the accuracy of the InternVL3_5-8B model, this
PR adds the corresponding configuration file to the CI. We need to add
the `-hf` suffix to avoid incompatibility with the `lm-eval`
preprocessor.

### How was this patch tested?

`pytest -sv ./tests/e2e/models/test_lm_eval_correctness.py --config
./tests/e2e/models/configs/InternVL3_5-8B.yaml`


- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

Signed-off-by: gcanlin <canlinguosdu@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

accuracy-test enable all accuracy test for PR module:tests ready read for review ready-for-test start test by label for PR

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants