Skip to content

[Benchmark] Cleanup deprecated nightly benchmark and adjust the docstring for performance benchmark#25786

Merged
yeqcharlotte merged 12 commits intovllm-project:mainfrom
KuntaiDu:kuntai-remove-nightly-bench
Oct 30, 2025
Merged

[Benchmark] Cleanup deprecated nightly benchmark and adjust the docstring for performance benchmark#25786
yeqcharlotte merged 12 commits intovllm-project:mainfrom
KuntaiDu:kuntai-remove-nightly-bench

Conversation

@KuntaiDu
Copy link
Collaborator

@KuntaiDu KuntaiDu commented Sep 26, 2025

Purpose

Cleanup the nightly benchmark suite because of two reasons:

  • The codepath of comparing vLLM against other engines is no longer maintained.
  • The performance benchmarking CI is now triggered under Pytorch Hub instead of in vLLM's CI pipeline.

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
@mergify mergify bot added documentation Improvements or additions to documentation ci/build performance Performance-related issues labels Sep 26, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request effectively cleans up the deprecated nightly benchmark suite by removing numerous files and reorganizing the directory structure. The changes align well with the stated purpose. However, I've identified a few inconsistencies in the documentation that were likely overlooked during the refactoring. These include incorrect paths in commands, broken links, and descriptions of functionality that has been removed. I've provided specific comments to help resolve these documentation issues.

@KuntaiDu KuntaiDu requested a review from bigPYJ1151 September 26, 2025 21:46
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
@yeqcharlotte
Copy link
Collaborator

could you fix markdown precommit?

## Performance benchmark quick overview

**Benchmarking Coverage**: latency, throughput and fix-qps serving on A100 (the support for FP8 benchmark on H100 is coming!) and Intel® Xeon® Processors, with different models.
**Benchmarking Coverage**: latency, throughput and fix-qps serving on A100, H100 and Intel® Xeon® Processors, with different models.
Copy link
Collaborator

@yeqcharlotte yeqcharlotte Sep 27, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we have b200 too~

not blocking this PR. cc: @huydhn @linzebing -- we need to add some new docs for vllm's continuous profiling and benchmarking runs! especially if these are locally runnable https://github.com/pytorch/pytorch-integration-testing/tree/main/vllm-benchmarks/benchmarks

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@namanlalitnyu : can you take this one?

Copy link
Contributor

@namanlalitnyu namanlalitnyu Sep 27, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure, I'll take this up.

But, I have a small concern with these changes, that we might have to consider. Here, we are migrating the file run-performance-benchmarks.sh from the nightly-benchmarks directory to the performance-benchmarks directory.
In order for our continuous vLLM benchmarking CI to work successfully in Pytorch-infra, I think we might have to update the vllm-benchmark.yml workflow as well, to adapt to these new changes, once they are merged. As, we would be referencing the incorrect directory location at a couple of places there.
This Draft PR (84) is ready and can be merged once these changes are verified from vLLM.

cc: @huydhn @linzebing

Copy link
Contributor

@namanlalitnyu namanlalitnyu Sep 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@yeqcharlotte @linzebing
Added the documentation for vLLM continuous profiling and benchmarking in this PR: 25819. But, I am not able to explicitly add reviewers to it 😅

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you @namanlalitnyu and @yeqcharlotte for checking. Please let me know if you prefer to put the renaming of nightly-benchmarks folder in a follow-up PR.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also B200 added.

Copy link
Collaborator

@yeqcharlotte yeqcharlotte left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks for cleaning this up! please fix some broken links and mkdown lints.

@huydhn
Copy link
Contributor

huydhn commented Sep 29, 2025

For the sake of completeness, here is a round of benchmark https://github.com/pytorch/pytorch-integration-testing/actions/runs/18112289981 running with the change from this PR

It looks like we need to tweak the workflows on PyTorch side a bit to handle the missing .buildkite/nightly-benchmarks/tests directory. Also, I don't think we can delete .buildkite/nightly-benchmarks/scripts/run-nightly-benchmarks.sh without moving that script to PyTorch and call the new script here https://github.com/pytorch/pytorch-integration-testing/blob/main/.github/workflows/vllm-benchmark.yml#L295

@mergify
Copy link

mergify bot commented Oct 3, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @KuntaiDu.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase label Oct 3, 2025
@mergify
Copy link

mergify bot commented Oct 8, 2025

Documentation preview: https://vllm--25786.org.readthedocs.build/en/25786/

@KuntaiDu
Copy link
Collaborator Author

KuntaiDu commented Oct 8, 2025

Thank you for all the feedbacks. I am travelling and will be back soon.

@KuntaiDu
Copy link
Collaborator Author

Now back to work

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
@mergify mergify bot removed the needs-rebase label Oct 24, 2025
Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
@KuntaiDu KuntaiDu requested a review from yeqcharlotte October 24, 2025 21:32
@KuntaiDu
Copy link
Collaborator Author

@yeqcharlotte I have updated this PR, would be really good if you can take a look again.

@yeqcharlotte
Copy link
Collaborator

@KuntaiDu fix the markdown lints?

Error: .buildkite/performance-benchmarks/README.md:8 MD012/no-multiple-blanks Multiple consecutive blank lines [Expected: 1; Actual: 2]
Error: .buildkite/performance-benchmarks/README.md:19 MD012/no-multiple-blanks Multiple consecutive blank lines [Expected: 1; Actual: 2]
Error: .buildkite/performance-benchmarks/README.md:137 MD012/no-multiple-blanks Multiple consecutive blank lines [Expected: 1; Actual: 2]

Copy link
Collaborator

@yeqcharlotte yeqcharlotte left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks for cleaning this up!

@yeqcharlotte yeqcharlotte enabled auto-merge (squash) October 28, 2025 03:19
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 28, 2025
@yeqcharlotte yeqcharlotte merged commit 8bff831 into vllm-project:main Oct 30, 2025
19 checks passed
@KuntaiDu KuntaiDu deleted the kuntai-remove-nightly-bench branch October 30, 2025 20:05
@KuntaiDu
Copy link
Collaborator Author

@namanlalitnyu It would be nice if we could also start the process of changing the pytorch CI correspondingly (like this draft PR (pytorch/pytorch-integration-testing#84))

@namanlalitnyu
Copy link
Contributor

@namanlalitnyu It would be nice if we could also start the process of changing the pytorch CI correspondingly (like this draft PR (pytorch/pytorch-integration-testing#84))
@KuntaiDu thanks for updating and the heads up. Yes, we have trigged the CI pipeline with these changes, and once they pass successfully, we will merge the updates there as well.

ilmarkov pushed a commit to neuralmagic/vllm that referenced this pull request Nov 7, 2025
…ring for performance benchmark (vllm-project#25786)

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
ZhengHongming888 pushed a commit to ZhengHongming888/vllm that referenced this pull request Nov 8, 2025
…ring for performance benchmark (vllm-project#25786)

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
rtourgeman pushed a commit to rtourgeman/vllm that referenced this pull request Nov 10, 2025
…ring for performance benchmark (vllm-project#25786)

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
…ring for performance benchmark (vllm-project#25786)

Signed-off-by: KuntaiDu <kuntai@uchicago.edu>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build documentation Improvements or additions to documentation performance Performance-related issues ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants