Skip to content

[Bug] Revert torch warning fix#31585

Merged
yewentao256 merged 3 commits intomainfrom
wentao-revert-torch-warning
Jan 5, 2026
Merged

[Bug] Revert torch warning fix#31585
yewentao256 merged 3 commits intomainfrom
wentao-revert-torch-warning

Conversation

@yewentao256
Copy link
Copy Markdown
Member

@yewentao256 yewentao256 commented Dec 31, 2025

Purpose

Fixes #31579

In 2.9.0, there was a warning in https://github.com/pytorch/pytorch/blob/0fabc3ba44823f257e70ce397d989c8de5e362c1/aten/src/ATen/Context.cpp#L79
But seems already be removed in PyTorch 2.9.1.

I think we can revert the fix now

Signed-off-by: yewentao256 <zhyanwentao@126.com>
@yewentao256 yewentao256 added ready ONLY add when PR is ready to merge/full CI is needed bug Something isn't working labels Dec 31, 2025
@mergify mergify bot added the v1 label Dec 31, 2025
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly updates the code to use torch.set_float32_matmul_precision instead of the deprecated torch.backends.cuda.matmul.fp32_precision. The changes in the modified files are accurate. However, the fix is incomplete as another usage of the deprecated API was missed in a different file. I've added a specific comment pointing out the location for a complete fix. Additionally, the pull request title and description are misleading and should be updated to reflect that this change applies a fix, rather than reverting one.

@yewentao256 yewentao256 enabled auto-merge (squash) January 1, 2026 00:48
@yewentao256 yewentao256 merged commit af9a7ec into main Jan 5, 2026
52 checks passed
@yewentao256 yewentao256 deleted the wentao-revert-torch-warning branch January 5, 2026 22:31
LucasWilkinson pushed a commit to neuralmagic/vllm that referenced this pull request Jan 6, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
yugong333 pushed a commit to yugong333/vllm that referenced this pull request Jan 9, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
akh64bit pushed a commit to akh64bit/vllm that referenced this pull request Jan 16, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
npanpaliya pushed a commit to odh-on-pz/vllm-cpu that referenced this pull request Feb 16, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
npanpaliya pushed a commit to odh-on-pz/vllm-cpu that referenced this pull request Feb 16, 2026
- [Misc] Implement `TokenizerLike.convert_tokens_to_ids`
(vllm-project/vllm#31796)
  [INFERENG-4151](https://issues.redhat.com/browse/INFERENG-4151)
- [Bug] Revert torch warning fix (vllm-project/vllm#31585)
  [INFERENG-4152](https://issues.redhat.com/browse/INFERENG-4152)
- [Bug] Fix AttributeError: `ColumnParallelLinear` object has no
attribute `weight_scale_inv` (vllm-project/vllm#30823)
  [INFERENG-4153](https://issues.redhat.com/browse/INFERENG-4153)
- Avoid `opencv-python-headless==4.13.0.90`, it's broken. See
opencv/opencv-python#1183
- [Bugfix] Handle mistral tokenizer in get_hf_processor
(vllm-project/vllm#31817)
  [INFERENG-4151](https://issues.redhat.com/browse/INFERENG-4151)
- [Bugfix] Fix Whisper/encoder-decoder GPU memory leak
vllm-project/vllm#32789
- [Model] Handle `trust_remote_code` for transformers backend
(vllm-project/vllm#32194) (fixes
GHSA-2pc9-4j83-qjmr)
- [Bugfix] CUDA: fix segfault by bumping numba to `numba==0.63.1`
([AIPCC-9384](https://issues.redhat.com/browse/AIPCC-9384))
- [Bugfix] pin `mistral_common==1.8.5` to avoid crash with Voxtral
([INFERENG-4154](https://issues.redhat.com/browse/INFERENG-4154))
- [Bugfix] fix tokenizer loading for mistral models
(vllm-project/vllm#33175)
  [INFERENG-4151](https://issues.redhat.com/browse/INFERENG-4151)
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
Signed-off-by: yewentao256 <zhyanwentao@126.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug Something isn't working ready ONLY add when PR is ready to merge/full CI is needed v1

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: VLLM_FLOAT32_MATMUL_PRECISION=tf32 does not set cublas tf32 matmul

2 participants