[DOC]Fix model weight download links#5436
Conversation
Updated download links for DeepSeek-V3.2 model weights. Signed-off-by: cookieyyds <126683903+cookieyyds@users.noreply.github.com>
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
There was a problem hiding this comment.
Code Review
This pull request updates a download link for the DeepSeek-V3.2 model weights. While the link update is correct, it introduces an inconsistency in the documentation. The model path used in the command examples throughout the file has not been updated to reflect the new model location, which will cause the tutorial steps to fail. I've added a critical comment to address this.
| - `DeepSeek-V3.2-Exp-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://modelers.cn/models/Modelers_Park/DeepSeek-V3.2-Exp-w8a8) | ||
| - `DeepSeek-V3.2`(BF16 version): require 2 Atlas 800 A3 (64G × 16) nodes or 4 Atlas 800 A2 (64G × 8) nodes. Model weight in BF16 not found now. | ||
| - `DeepSeek-V3.2-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://modelers.cn/models/Eco-Tech/DeepSeek-V3.2-w8a8-mtp-QuaRot) | ||
| - `DeepSeek-V3.2-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://www.modelscope.cn/models/vllm-ascend/DeepSeek-V3.2-W8A8/) |
There was a problem hiding this comment.
While the download link has been updated, the model path used in the command examples throughout this document still refers to the old model path. This will cause the commands to fail for users who download the model from the new link.
The old path was /root/.cache/Eco-Tech/DeepSeek-V3.2-w8a8-mtp-QuaRot.
The new path should likely be /root/.cache/vllm-ascend/DeepSeek-V3.2-W8A8 to match the new ModelScope URL.
Please update all occurrences of the old path in this file. The affected lines are:
- Line 292
- Line 367
- Line 444
- Line 523
- Line 631
- Line 669
| - `DeepSeek-V3.2-Exp-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://modelers.cn/models/Modelers_Park/DeepSeek-V3.2-Exp-w8a8) | ||
| - `DeepSeek-V3.2`(BF16 version): require 2 Atlas 800 A3 (64G × 16) nodes or 4 Atlas 800 A2 (64G × 8) nodes. Model weight in BF16 not found now. | ||
| - `DeepSeek-V3.2-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://modelers.cn/models/Eco-Tech/DeepSeek-V3.2-w8a8-mtp-QuaRot) | ||
| - `DeepSeek-V3.2-w8a8`(Quantized version): require 1 Atlas 800 A3 (64G × 16) node or 2 Atlas 800 A2 (64G × 8) nodes. [Download model weight](https://www.modelscope.cn/models/vllm-ascend/DeepSeek-V3.2-W8A8/) |
There was a problem hiding this comment.
@zhangxinyuehfad Let's align the model name on modelscope and the source model later, just to make sure it doesn't confuse users
…to eplb_refactor * 'main' of https://github.com/vllm-project/vllm-ascend: (46 commits) [Feature] Support to use fullgraph with eagle (vllm-project#5118) [EPLB][refactor] Modification of the initialization logic for expert_map and log2phy(depend on pr5285) (vllm-project#5311) [Refactor]6/N Extract common code of class AscendMLAImpl (vllm-project#5314) [Refactor] cache cos/sin in mla & remove parameter model in builder. (vllm-project#5277) update vllm pin to 12.27 (vllm-project#5412) [ReleaseNote] Add release note for v0.13.0rc1 (vllm-project#5334) [Bugfix] Correctly handle the output shape in multimodal attention (vllm-project#5443) Fix nightly (vllm-project#5413) [bugfix] fix typo of _skip_all_reduce_across_dp_group (vllm-project#5435) [Doc]modify pcp tutorial doc (vllm-project#5440) [Misc] fast fail for exiting if tools/install_flash_infer_attention_score_ops_a2.sh (vllm-project#5422) [Doc] Update DeepSeek V3.1/R1 2P1D doc (vllm-project#5387) [DOC]Fix model weight download links (vllm-project#5436) [Doc] Modify DeepSeek-R1/V3.1 documentation (vllm-project#5426) Revert "[feat] enable hierarchical mc2 ops on A2 by default (vllm-project#5300)" (vllm-project#5434) [Bugfix] fix greedy temperature detection (vllm-project#5417) [doc] Update Qwen3-235B doc for reproducing latest performance (vllm-project#5323) [feat] enable hierarchical mc2 ops on A2 by default (vllm-project#5300) [Doc] delete environment variable HCCL_OP_EXPANSION_MODE in DeepSeekV3.1/R1 (vllm-project#5419) [Doc] add long_sequence feature user guide (vllm-project#5343) ...
Updated download links for DeepSeek-V3.2 model weights. - vLLM version: release/v0.13.0 - vLLM main: vllm-project/vllm@81786c8 Signed-off-by: cookieyyds <126683903+cookieyyds@users.noreply.github.com> Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
Updated download links for DeepSeek-V3.2 model weights. - vLLM version: release/v0.13.0 - vLLM main: vllm-project/vllm@81786c8 Signed-off-by: cookieyyds <126683903+cookieyyds@users.noreply.github.com>
Updated download links for DeepSeek-V3.2 model weights. - vLLM version: release/v0.13.0 - vLLM main: vllm-project/vllm@81786c8 Signed-off-by: cookieyyds <126683903+cookieyyds@users.noreply.github.com> Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
Updated download links for DeepSeek-V3.2 model weights.