Skip to content

Clear torch.compile cache before each benchmark#13423

Closed
huydhn wants to merge 1 commit intovllm-project:mainfrom
huydhn:cleanup-cache-before-benchmark
Closed

Clear torch.compile cache before each benchmark#13423
huydhn wants to merge 1 commit intovllm-project:mainfrom
huydhn:cleanup-cache-before-benchmark

Conversation

@huydhn
Copy link
Contributor

@huydhn huydhn commented Feb 17, 2025

This is a work around that I'm using to run v1 benchmark due to #13392

cc @youkaichao @ywang96

@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@mergify mergify bot added the ci/build label Feb 17, 2025
Copy link
Member

@ywang96 ywang96 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey @huydhn - Thanks for the PR! I don't think this PR is needed now since we're going to merge #13445. Can you run the benchmark again and see if you're still seeing the same issue?

@huydhn
Copy link
Contributor Author

huydhn commented Feb 18, 2025

Hey @huydhn - Thanks for the PR! I don't think this PR is needed now since we're going to merge #13445. Can you run the benchmark again and see if you're still seeing the same issue?

Sounds good, let me try #13445.

This change might make sense if we want all benchmark runs to be the same (with cold start), but I guess this is a small point.

@huydhn huydhn closed this Feb 18, 2025
@youkaichao
Copy link
Member

This change might make sense if we want all benchmark runs to be the same (with cold start), but I guess this is a small point.

vLLM guarantees that all compilation finishes in its own warmup phase. no compilation will happen when we serve requests.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants