[Docs] Add InternVL series tutorial for single NPU#3664
[Docs] Add InternVL series tutorial for single NPU#3664gcanlin wants to merge 4 commits intovllm-project:mainfrom
Conversation
Signed-off-by: gcanlin <canlinguosdu@gmail.com>
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
There was a problem hiding this comment.
Code Review
This pull request adds a new tutorial for running InternVL series models on a single NPU. The tutorial is well-structured and provides clear instructions for both offline inference and online serving. However, there is a significant issue with the model path used in the examples, which points to a non-public model. This will prevent users from following the tutorial. My review includes a comment to address this.
|
Please add an accuracy test with same parameter with doc before merge. |
|
This pull request has conflicts, please resolve those before we can evaluate the pull request. |
### What this PR does / why we need it? As a validation for #3664, add end-to-end tests to monitor the InternVL model and ensure its continuous proper operation. This PR is only for single-card. So the models that have more parameters than 8B like 78B are needed to test using multi-cards. ### Does this PR introduce _any_ user-facing change? None. ### How was this patch tested? `pytest -sv tests/e2e/singlecard/multi-modal/test_internvl.py` - vLLM version: v0.11.0 - vLLM main: vllm-project/vllm@83f478b --------- Signed-off-by: gcanlin <canlinguosdu@gmail.com>
| ] | ||
|
|
||
| # Same image for all prompts | ||
| images = [image] * len(prompts) |
There was a problem hiding this comment.
This images variable is not used, should it be deleted?
There was a problem hiding this comment.
Sorry for the late response. It seems that I lost some code here. I will complete it later. Thanks!
### What this PR does / why we need it? As a validation for vllm-project#3664, add end-to-end tests to monitor the InternVL model and ensure its continuous proper operation. This PR is only for single-card. So the models that have more parameters than 8B like 78B are needed to test using multi-cards. ### Does this PR introduce _any_ user-facing change? None. ### How was this patch tested? `pytest -sv tests/e2e/singlecard/multi-modal/test_internvl.py` - vLLM version: v0.11.0 - vLLM main: vllm-project/vllm@83f478b --------- Signed-off-by: gcanlin <canlinguosdu@gmail.com> Signed-off-by: luolun <luolun1995@cmbchina.com>
### What this PR does / why we need it? As a validation for vllm-project#3664, add end-to-end tests to monitor the InternVL model and ensure its continuous proper operation. This PR is only for single-card. So the models that have more parameters than 8B like 78B are needed to test using multi-cards. ### Does this PR introduce _any_ user-facing change? None. ### How was this patch tested? `pytest -sv tests/e2e/singlecard/multi-modal/test_internvl.py` - vLLM version: v0.11.0 - vLLM main: vllm-project/vllm@83f478b --------- Signed-off-by: gcanlin <canlinguosdu@gmail.com> Signed-off-by: hwhaokun <haokun0405@163.com>
### What this PR does / why we need it? As a validation for vllm-project#3664, add end-to-end tests to monitor the InternVL model and ensure its continuous proper operation. This PR is only for single-card. So the models that have more parameters than 8B like 78B are needed to test using multi-cards. ### Does this PR introduce _any_ user-facing change? None. ### How was this patch tested? `pytest -sv tests/e2e/singlecard/multi-modal/test_internvl.py` - vLLM version: v0.11.0 - vLLM main: vllm-project/vllm@83f478b --------- Signed-off-by: gcanlin <canlinguosdu@gmail.com> Signed-off-by: nsdie <yeyifan@huawei.com>
### What this PR does / why we need it? As a validation for vllm-project#3664, add end-to-end tests to monitor the InternVL model and ensure its continuous proper operation. This PR is only for single-card. So the models that have more parameters than 8B like 78B are needed to test using multi-cards. ### Does this PR introduce _any_ user-facing change? None. ### How was this patch tested? `pytest -sv tests/e2e/singlecard/multi-modal/test_internvl.py` - vLLM version: v0.11.0 - vLLM main: vllm-project/vllm@83f478b --------- Signed-off-by: gcanlin <canlinguosdu@gmail.com>
|
Any progress? If this PR is still alive, please rebase to main and make CI happy. Thanks |
YES, I will refactor this PR later, maybe will need about 2 days. |
|
Any progress? |
|
Please rebase and fix the merge conflicts |
|
no update for long time, close this now. Feel free to reopen if it's still needed. |
What this PR does / why we need it?
Closes #3508.
Does this PR introduce any user-facing change?
No.
How was this patch tested?
No need to test. All scripts and Python code in the docs has been tested.