-
Notifications
You must be signed in to change notification settings - Fork 54
Feature: cuda specific make_attribute_wrapper #193
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
gmarkall
merged 6 commits into
NVIDIA:main
from
ZzEeKkAa:yhavrylko/feature/make_attribute_wrapper
Apr 30, 2025
Merged
Feature: cuda specific make_attribute_wrapper #193
gmarkall
merged 6 commits into
NVIDIA:main
from
ZzEeKkAa:yhavrylko/feature/make_attribute_wrapper
Apr 30, 2025
+65
−5
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
gmarkall
requested changes
Apr 14, 2025
Contributor
gmarkall
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the code looks good (a couple of small comments on the diff). I think we also need:
- A merge of
mainto fix CI. - A test of the functionality it does enable - I can see it doesn't break any of the existing use cases in the test suite, but I assume they all used models that were not only in the CUDA data model manager. Can we have a test using a model that's only present in the CUDA data model manager?
- If this is not the appropriate test, then I have misunderstood the purpose of the PR, and a clarification and appropriate test would also be appreciated 🙂
Co-authored-by: Graham Markall <535640+gmarkall@users.noreply.github.com>
…e_attribute_wrapper
ZzEeKkAa
commented
Apr 16, 2025
gmarkall
approved these changes
Apr 30, 2025
gmarkall
added a commit
to gmarkall/numba-cuda
that referenced
this pull request
May 2, 2025
- Fix Invalid NVVM IR emitted when lowering shfl_sync APIs (NVIDIA#231) - Add Bfloat16 Low++ Bindings (NVIDIA#166) - Fix cuda.jit decorator inline (NVIDIA#181) - Feature: cuda specific make_attribute_wrapper (NVIDIA#193) - return a none tuple if no libdevice path is found (NVIDIA#234)
Closed
gmarkall
added a commit
to gmarkall/numba-cuda
that referenced
this pull request
May 3, 2025
- Local variable debug info deduplication (NVIDIA#222) - Fix package installation for wheels CI (NVIDIA#238) - Fix Invalid NVVM IR emitted when lowering shfl_sync APIs (NVIDIA#231) - Add Bfloat16 Low++ Bindings (NVIDIA#166) - Fix cuda.jit decorator inline (NVIDIA#181) - Feature: cuda specific make_attribute_wrapper (NVIDIA#193) - return a none tuple if no libdevice path is found (NVIDIA#234)
Merged
gmarkall
added a commit
that referenced
this pull request
May 3, 2025
- Local variable debug info deduplication (#222) - Fix package installation for wheels CI (#238) - Fix Invalid NVVM IR emitted when lowering shfl_sync APIs (#231) - Add Bfloat16 Low++ Bindings (#166) - Fix cuda.jit decorator inline (#181) - Feature: cuda specific make_attribute_wrapper (#193) - return a none tuple if no libdevice path is found (#234)
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Vendor numba's
make_attribute_wrapperuntil it supportscudatarget. Vendored function was changed to use cuda specific data model manager chained with a default data model manager. That makes it possible to wrap array attributes that are not supported bynumba, but supported bynumba-cuda(likefp16arrays).