Use standalone_compile by default in torch >= 2.8.0#18846
Use standalone_compile by default in torch >= 2.8.0#18846houseroad merged 1 commit intovllm-project:mainfrom
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
vllm/envs.py
Outdated
There was a problem hiding this comment.
Wondering why there is TEST in the name?
Also do we want to turn on the flag by default?
There was a problem hiding this comment.
The flag was already named VLLM_TEST_STANDALONE_COMPILE in a previous PR. I can rename it if you want, to something like VLLM_USE_STANDALONE_COMPILE, but I was following the naming of VLLM_TEST_DYNAMO_FULLGRAPH_CAPTURE.
This PR turns the flag on by default.
There was a problem hiding this comment.
Yeah, let's use something like VLLM_USE_STANDALONE_COMPILE? I feel VLLM_TEST_STANDALONE_COMPILE is misleading, like some flag used in the test only.
0c6e52b to
d8e7241
Compare
houseroad
left a comment
There was a problem hiding this comment.
Looks good, thanks for updating the env var name.
vllm/envs.py
Outdated
There was a problem hiding this comment.
Does internal mean meta internal? If so, it's not internal anymore :-)
There was a problem hiding this comment.
internal does not mean "meta internal", it means "not public API". I'll drop the "internal" word.
d8e7241 to
e239a1a
Compare
This includes the current PyTorch nightlies. It also renames the VLLM_TEST_STANDALONE_COMPILE envvar to VLLM_USE_STANDALONE_COMPILE to make it clearer. Test Plan: - in vllm-project#17057, I verified that running https://gist.github.com/zou3519/aebb622714e80f4cd4c369472f2372cd with or without VLLM_TEST_STANDALONE_COMPILE resulted in Inductor producing the same exact output code (via tlparse). I did this for the cold-start case and the warm start case. - there are vllm x torch nightly tests in CI that I will trigger on this PR. Signed-off-by: rzou <zou3519@gmail.com>
e239a1a to
ec26191
Compare
|
Cc @jerryzh168 we shud make sure that we don't land the change disabling compile cache for 2.8+ |
|
Btw, I thought I changed the commit message but looks like I forgot to or it didn't update the PR body. The envvar was renamed to VLLM_USE_STANDALONE_COMPILE in this PR. |
Signed-off-by: rzou <zou3519@gmail.com> Signed-off-by: amit <amit.man@gmail.com>
Signed-off-by: rzou <zou3519@gmail.com> Signed-off-by: amit <amit.man@gmail.com>
Signed-off-by: rzou <zou3519@gmail.com>
* use 2025.1.1 instead (vllm-project#196) Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> * Use standalone_compile by default in torch >= 2.8.0 (vllm-project#18846) Signed-off-by: rzou <zou3519@gmail.com> * fix xpu compile issue --------- Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> Signed-off-by: rzou <zou3519@gmail.com> Co-authored-by: Richard Zou <zou3519@users.noreply.github.com>
* use 2025.1.1 instead (vllm-project#196) Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> * Use standalone_compile by default in torch >= 2.8.0 (vllm-project#18846) Signed-off-by: rzou <zou3519@gmail.com> * fix xpu compile issue --------- Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> Signed-off-by: rzou <zou3519@gmail.com> Co-authored-by: Richard Zou <zou3519@users.noreply.github.com>
* use 2025.1.1 instead (vllm-project#196) Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> * Use standalone_compile by default in torch >= 2.8.0 (vllm-project#18846) Signed-off-by: rzou <zou3519@gmail.com> * fix xpu compile issue --------- Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> Signed-off-by: rzou <zou3519@gmail.com> Co-authored-by: Richard Zou <zou3519@users.noreply.github.com>
* use 2025.1.1 instead (vllm-project#196) Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> * Use standalone_compile by default in torch >= 2.8.0 (vllm-project#18846) Signed-off-by: rzou <zou3519@gmail.com> * fix xpu compile issue --------- Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> Signed-off-by: rzou <zou3519@gmail.com> Co-authored-by: Richard Zou <zou3519@users.noreply.github.com>
* use 2025.1.1 instead (vllm-project#196) Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> * Use standalone_compile by default in torch >= 2.8.0 (vllm-project#18846) Signed-off-by: rzou <zou3519@gmail.com> * fix xpu compile issue --------- Signed-off-by: Kunshang Ji <kunshang.ji@intel.com> Signed-off-by: rzou <zou3519@gmail.com> Co-authored-by: Richard Zou <zou3519@users.noreply.github.com>
This includes the current PyTorch nightlies.
Test Plan: