[Build] Allow shipping PTX on a per-file basis#18155
[Build] Allow shipping PTX on a per-file basis#18155simon-mo merged 7 commits intovllm-project:mainfrom
Conversation
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
| endif() | ||
|
|
||
| list(SORT SRC_CUDA_ARCHS COMPARE NATURAL ORDER ASCENDING) | ||
| list(SORT _SRC_CUDA_ARCHS COMPARE NATURAL ORDER ASCENDING) |
There was a problem hiding this comment.
Do the _TGT_CUDA_ARCHS need to be sorted too?
There was a problem hiding this comment.
for a more general utility you're right they should be!; but the target arches come from extract_unique_cuda_archs_ascending so they are already sorted. I can open up a now PR to refactor some of this though; kinda want to preserve current behavior as much as possible in this one
bnellnm
left a comment
There was a problem hiding this comment.
Looks reasonable to me.
There was a problem hiding this comment.
Seems relatively safe to me. There might be a regression for marlin because of slow bf16 convert on A100 (IIRC) that might transfer to newer hardware, but also might not. Ultimately shouldn't be that big of a deal. @jinzhen-lin please step in if you have concerns with this since you refactored marlin most recently.
mgoin
left a comment
There was a problem hiding this comment.
The failures look closely related
Currently, I believe that the performance of |
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com>
|
Failures resolved |
mgoin
left a comment
There was a problem hiding this comment.
This is a nice middle ground, thanks for getting it working!
Signed-off-by: Lucas Wilkinson <lwilkinson@neuralmagic.com> Signed-off-by: Yuqi Zhang <yuqizhang@google.com>
323.20 MB -> 324.36 MB
To help with the growing wheel size due to Blackwell allow for shipping PTX for heavy kernels that don't take advantage of new hardware features. Theres enough different gencodes now for certain kernels it makes sense to ship a single PTX implementation instead of multiple SASS. Currently mildly grows the wheel size but should help keep it capped as the Blackwell gencodes are added