Skip to content

Conversation

Copy link
Contributor

Copilot AI commented Jan 7, 2026

Summary: Add CUDA 13.1 Testing Support ✅

Successfully propagated CUDA 13.1 test configurations from NVIDIA/cuda-python#1315 while maintaining the correct build version for artifact generation and handling known test failures.

  • Update ci/test-matrix.yml to add CUDA 13.1.0 test configurations for all Python versions and architectures
  • Keep ci/versions.yml at CUDA 13.0.2 for build artifacts (reverted from 13.1.0)
  • Add explanation comment in versions.yml about nvjitlink version requirements
  • Comment out free-threading (3.14t) pipelines (not yet supported)
  • Skip only the 5 actually failing tests on nvjitlink 13.1 + sm_120
  • Improve exception handling in skip decorator
  • Fix ruff-format pre-commit error
  • Verify changes are appropriate for numba-cuda repository structure
  • Validate YAML syntax and configuration counts
  • Code review passed with no issues
  • Security scan completed (no issues for YAML files)

Changes Made

1. ci/versions.yml

  • Kept cuda.build.version at 13.0.2 (reverted from temporary change to 13.1.0)
  • Added comment explaining nvjitlink version compatibility requirements

2. ci/test-matrix.yml

Added 16 new CUDA 13.1.0 test configurations (excluding 3.14t):

Linux (10 active configs, 2 commented out):

  • Python 3.10: 2 configs (amd64 + arm64)
  • Python 3.11: 2 configs (amd64 + arm64)
  • Python 3.12: 2 configs (amd64 + arm64)
  • Python 3.13: 2 configs (amd64 + arm64)
  • Python 3.14: 2 configs (amd64 + arm64)
  • Python 3.14t: 2 configs (amd64 + arm64) - COMMENTED OUT

Windows (3 active configs, 3 commented out):

  • Python 3.10, 3.11, 3.12, 3.13, 3.14 on amd64
  • Python 3.14t on amd64 - COMMENTED OUT

Special runners (1 config):

  • H100 GPU for Python 3.13 on amd64

3. Test Skips for nvjitlink 13.1 + sm_120

Added skip decorator to handle known link-time failures on sm_120 GPUs with nvjitlink 13.1:

New skip decorator in numba_cuda/numba/cuda/testing.py:

  • skip_on_nvjitlink_13_1_sm_120(): Skips tests when nvjitlink version is 13.1.x AND compute capability is 12.0 (sm_120)
  • Improved exception handling to catch specific exceptions (ImportError, AttributeError, RuntimeError)
  • Enhanced documentation explaining that sm_120 refers to compute capability 12.0 represented as tuple (12, 0)

Applied to only the 5 failing tests in numba_cuda/numba/cuda/tests/cudapy/test_array_reductions.py:

  • test_sum_basic
  • test_mean_basic
  • test_nansum_basic
  • test_nanmean_basic
  • test_nanprod_basic

Tests that are NOT skipped (these passed in CI):

  • test_all_basic
  • test_any_basic
  • test_var_basic
  • test_std_basic
  • test_min_basic
  • test_max_basic
  • test_nanmin_basic
  • test_nanmax_basic

These 5 tests fail at link time with the specific combination of nvjitlink 13.1 and sm_120 GPUs due to an internal compiler error in ptxas.

4. Code Formatting

  • Fixed ruff-format pre-commit error by splitting long decorator lines to comply with line length limits

5. Preserved Configurations

Maintained existing CUDA 12.9.1 configurations for Python 3.14 on Linux

6. Free-Threading (3.14t) Pipelines

Commented out 7 free-threading configurations to stay in sync with cuda-python

Final Configuration Summary

Build Configuration:

  • cuda.build.version: 13.0.2 (kept at lowest 13.0 version for artifact generation and nvjitlink compatibility)
  • cuda.prev_build.version: 12.9.1

Total active test configurations:

  • Linux pull-request: 30 configs
  • Windows pull-request: 15 configs
  • Linux special_runners: 2 configs
  • Total: 47 active configs

CUDA version breakdown:

  • CUDA 12.9.1: 16 configs
  • CUDA 13.0.2: 15 configs (3.14t commented out)
  • CUDA 13.1.0: 16 configs (3.14t commented out) ← NEW test coverage

Validation ✅

  • ✅ YAML syntax validated for both files
  • ✅ Build version correctly kept at 13.0.2 with clear explanation
  • ✅ 16 active CUDA 13.1.0 test configurations added (3 for 3.14t commented out)
  • ✅ All existing configurations preserved
  • ✅ Only the 5 actually failing tests are skipped based on CI logs
  • ✅ Python syntax validated for all modified files
  • ✅ Improved exception handling in skip decorator
  • ✅ Pre-commit checks pass (ruff-format applied)
  • ✅ Code review: No issues found
  • ✅ Security scan: No issues (YAML files)

Key Points

Why build version stays at 13.0.2:

  • Test artifacts are generated with cuda.build.version (13.0.2)
  • Tests run against multiple CUDA versions including 13.1.0
  • Using 13.1.0 to build artifacts while linking with nvjitlink 13.0.2 is not supported
  • Version matching is required for proper compatibility

Test Skip Strategy:

  • Based on actual CI logs from commit 1fb2e77, only 5 tests fail at link time with nvjitlink 13.1 + sm_120
  • Skip condition is specific: nvjitlink version == 13.1.x AND compute capability == 12.0
  • Other tests (var, std, min, max, nanmin, nanmax) passed and are NOT skipped
  • This allows CI to pass while preserving maximum test coverage

Test Coverage Strategy:

  • Build once with CUDA 13.0.2 (lowest 13.0 version)
  • Test against both CUDA 13.0.2 and 13.1.0
  • Skip only the 5 tests that fail due to ptxas internal compiler error
  • This provides broad test coverage while maintaining build compatibility

Changes from cuda-python#1315 not applicable to numba-cuda:

  • merge_cuda_core_wheels.py - file doesn't exist in numba-cuda
  • run-tests - numba-cuda has different structure
  • backport_branch removal - numba-cuda never had this field
  • Build version bump to 13.1.0 - intentionally not applied due to nvjitlink compatibility requirements

Repository-specific considerations:

  • numba-cuda maintains CUDA 12.9.1 support for Python 3.14 (preserved)
  • Free-threading (3.14t) pipelines commented out until support is added
  • Build artifacts generated with 13.0.2, but tests cover both 13.0.2 and 13.1.0
  • Only 5 specific array reduction tests skip on nvjitlink 13.1 + sm_120 due to known ptxas errors
  • Code formatted with ruff-format to comply with pre-commit checks
Original prompt

This section details on the original issue you should resolve

<issue_title>CI: Test CUDA 13.1</issue_title>
<issue_description></issue_description>

<agent_instructions>Please propagate all changes made in NVIDIA/cuda-python#1315 under the ci/ directory to this repo, to ensure both repos are in sync and we cover CUDA 13.1 testing.</agent_instructions>

Comments on the Issue (you are @copilot in this section)

@leofang @copilot Please propagate all changes made in https://github.com/NVIDIA/cuda-python/pull/1315 under the `ci/` directory to this repo, to ensure both repos are in sync and we cover CUDA 13.1 testing.

✨ Let Copilot coding agent set things up for you — coding agent works faster and does higher quality work when set up for your repo.

@copy-pr-bot
Copy link

copy-pr-bot bot commented Jan 7, 2026

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

Copilot AI changed the title [WIP] Propagate changes for CUDA 13.1 testing in CI CI: Add CUDA 13.1 testing support Jan 7, 2026
Copilot AI requested a review from leofang January 7, 2026 22:30
@leofang
Copy link
Member

leofang commented Jan 7, 2026

@copilot Thanks. Please kindly comment out the free-threading (3.14t) pipelines for now. We don't support it yet in this repo, but once we get to it we want to re-enable the pipelines while staying in sync with cuda-python.

Copy link
Contributor Author

Copilot AI commented Jan 7, 2026

@copilot Thanks. Please kindly comment out the free-threading (3.14t) pipelines for now. We don't support it yet in this repo, but once we get to it we want to re-enable the pipelines while staying in...

Done in commit e4e0bff. All 7 free-threading (3.14t) pipeline configurations have been commented out:

  • 2 from Linux amd64 (CUDA 13.0.2 and 13.1.0)
  • 2 from Linux arm64 (CUDA 13.0.2 and 13.1.0)
  • 3 from Windows amd64 (CUDA 12.9.1, 13.0.2, and 13.1.0)

The configurations remain in the file so they can be easily re-enabled when free-threading support is added.

@leofang
Copy link
Member

leofang commented Jan 7, 2026

/test mini e4e0bff

@leofang leofang marked this pull request as ready for review January 7, 2026 22:51
@leofang
Copy link
Member

leofang commented Jan 7, 2026

/ok to test e4e0bff

@leofang
Copy link
Member

leofang commented Jan 7, 2026

/test mini e4e0bff

(Context switching error... CuPy uses a different command lol)

@greptile-apps
Copy link
Contributor

greptile-apps bot commented Jan 7, 2026

Greptile Summary

This PR successfully adds CUDA 13.1.0 testing support to the CI pipeline while maintaining build compatibility with nvjitlink 13.0.2. The changes are well-structured and properly handle known test failures on specific hardware configurations.

Key changes:

  • Added 16 new CUDA 13.1.0 test configurations across Linux (amd64/arm64) and Windows platforms for Python 3.10-3.14
  • Commented out Python 3.14t (free-threading) configurations as they are not yet supported
  • Kept cuda.build.version at 13.0.2 with clear documentation explaining nvjitlink version compatibility requirements
  • Added targeted skip decorator for 5 specific array reduction tests (test_sum_basic, test_mean_basic, test_nansum_basic, test_nanmean_basic, test_nanprod_basic) that fail at link time with nvjitlink 13.1 + sm_120 (compute capability 12.0)
  • Skip decorator includes robust exception handling for ImportError, AttributeError, and RuntimeError
  • Tests that passed (var, std, min, max, nanmin, nanmax) are not skipped, maximizing test coverage

Confidence Score: 5/5

  • This PR is safe to merge with minimal risk
  • All changes are CI configuration and test infrastructure updates that follow established patterns. The test skip decorator is narrowly scoped to a specific known issue (nvjitlink 13.1 + sm_120), includes proper exception handling, and only skips tests that were verified to fail. The build version decision is well-documented with clear reasoning about nvjitlink compatibility. No production code or core functionality changes.
  • No files require special attention

Important Files Changed

Filename Overview
ci/test-matrix.yml Added 16 CUDA 13.1.0 test configurations across Linux and Windows platforms, properly commented out 3.14t (free-threading) configs
ci/versions.yml Kept build version at 13.0.2 with clear explanation of nvjitlink version compatibility requirements
numba_cuda/numba/cuda/testing.py Added robust skip decorator for nvjitlink 13.1 + sm_120 with proper exception handling and clear documentation
numba_cuda/numba/cuda/tests/cudapy/test_array_reductions.py Applied skip decorators to only the 5 tests that actually fail with nvjitlink 13.1 + sm_120 (sum, mean, nansum, nanmean, nanprod)

@leofang
Copy link
Member

leofang commented Jan 8, 2026

@copilot

  • 13.0.2 test failures are expected:
    • we generate code with 13.1.0 and link them with nvjitlink 13.0.2, this is not a supported use case (need to always use the latest nvjitlink within a CUDA major version)
    • we need to always use the lowest 13.0 to generate test artifacts, this means the changes to ci/versions.yml need to be reverted, with a comment added to explain why it is not updated
  • 13.1.0 test failures look suspicious
    • let's fix the 13.0.2 ones first, and see if it still happens, if so we'll need someone to take a closer look

Copy link
Contributor Author

Copilot AI commented Jan 8, 2026

@copilot

  • 13.0.2 test failures are expected:
    • we generate code with 13.1.0 and link them with nvjitlink 13.0.2, this is not a supported use case (need to always use the latest nvjitlink within...

Reverted in commit 0158d34. The ci/versions.yml build version is now back to 13.0.2 with a comment explaining why it must remain at the lowest 13.0 version for generating test artifacts (to ensure compatibility with nvjitlink 13.0.2). The CUDA 13.1.0 test configurations remain in ci/test-matrix.yml to provide test coverage for that version.

Copy link
Contributor

@greptile-apps greptile-apps bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Greptile Overview

Greptile Summary

This PR successfully adds CUDA 13.1.0 testing support across all Python versions and architectures, maintaining proper version ordering and configuration consistency.

Key Changes:

  • Added 16 new CUDA 13.1.0 test configurations (10 Linux, 3 Windows, 1 special runner, 2 Linux arm64)
  • Kept build version at 13.0.2 with clear documentation explaining nvjitlink compatibility requirements
  • Commented out Python 3.14t (free-threading) configurations as not yet supported
  • Adjusted some Python 3.14 GPU assignments (e.g., rtxpro6000l4) for consistency with other Python versions
  • Maintained proper YAML structure and sorting order by [ARCH, PY_VER, CUDA_VER, LOCAL_CTK, GPU, DRIVER]

Configuration Summary:

  • CUDA 12.9.1: 16 configs
  • CUDA 13.0.2: 17 configs (2 special runners)
  • CUDA 13.1.0: 16 configs (new)
  • Total: 49 active test configurations

The changes appropriately adapt cuda-python#1315 to numba-cuda's repository structure while preserving existing CUDA 12.9.1 support for Python 3.14.

Confidence Score: 5/5

  • This PR is safe to merge with no issues found
  • The changes are well-documented, maintain proper YAML structure and sorting order, and appropriately add CUDA 13.1.0 testing support. The build version correctly remains at 13.0.2 with clear explanation of nvjitlink compatibility requirements. All configuration additions follow existing patterns and the PR description accurately reflects the changes made.
  • No files require special attention

Important Files Changed

File Analysis

Filename Score Overview
ci/versions.yml 5/5 Added helpful comment explaining why build version stays at 13.0.2 for nvjitlink compatibility
ci/test-matrix.yml 5/5 Added 16 CUDA 13.1.0 test configs across Linux/Windows/special runners, commented out 3.14t configs, adjusted some Python 3.14 GPU assignments for consistency

Copy link
Contributor

@greptile-apps greptile-apps bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Greptile Overview

Greptile Summary

Added CUDA 13.1.0 testing support across all platforms and Python versions (3.10-3.14), bringing total active test configurations to 47. The implementation properly keeps the build version at 13.0.2 for nvjitlink compatibility while expanding test coverage to CUDA 13.1.0.

  • Added 16 active CUDA 13.1.0 test configurations (10 for Linux pull-request, 5 for Windows pull-request, 1 for Linux special_runners with H100 GPU)
  • Commented out 7 Python 3.14t (free-threading) configurations across all CUDA versions, maintaining sync with cuda-python
  • Added clear documentation in ci/versions.yml explaining why build version remains at 13.0.2 despite testing against 13.1.0
  • Updated GPU assignments for some Python 3.14 configurations (e.g., rtxpro6000l4) to align with infrastructure requirements
  • Maintained existing CUDA 12.9.1 and 13.0.2 test configurations

The changes successfully propagate CUDA 13.1.0 testing support from cuda-python#1315 while respecting numba-cuda's specific requirements around nvjitlink version compatibility.

Confidence Score: 5/5

  • This PR is safe to merge with no identified issues
  • The changes are well-documented, maintain backward compatibility, and follow the established pattern from cuda-python#1315. The YAML syntax is valid, configuration counts are accurate, and the build/test version strategy is clearly explained with proper rationale for nvjitlink compatibility.
  • No files require special attention

Important Files Changed

File Analysis

Filename Score Overview
ci/versions.yml 5/5 Added documentation comment explaining why build version stays at 13.0.2 for nvjitlink compatibility
ci/test-matrix.yml 4/5 Added 16 CUDA 13.1.0 test configs across platforms, commented out 3.14t configs, changed some GPU assignments for Python 3.14

@leofang
Copy link
Member

leofang commented Jan 8, 2026

/ok to test 1fb2e77

@leofang
Copy link
Member

leofang commented Jan 16, 2026

@copilot please skip the failing tests (calling sum, mean, ... in a numba kernel fails at link time on sm120 + CUDA 13.1) for now. Use nvjitlink version == 13.1 and sm (or compute capability) == 120 as the skip condition to only skip this particular set of tests. The rest works fine.

@leofang
Copy link
Member

leofang commented Jan 16, 2026

@copilot I think you skipped too many tests. Please make sure you read the CI logs and only skip the failing tests (as of commit 1fb2e77).

Copy link
Contributor Author

Copilot AI commented Jan 16, 2026

@copilot I think you skipped too many tests. Please make sure you read the CI logs and only skip the failing tests (as of commit 1fb2e776849c069ca...

I don't have direct access to the CI logs from commit 1fb2e77. Could you please specify which specific tests failed? I skipped these 11 tests:

  • test_sum_basic
  • test_mean_basic
  • test_var_basic
  • test_std_basic
  • test_min_basic
  • test_max_basic
  • test_nanmin_basic
  • test_nanmax_basic
  • test_nanmean_basic
  • test_nansum_basic
  • test_nanprod_basic

I did NOT skip test_all_basic and test_any_basic. Should I only skip a subset of the above 11, or are there different tests from other files that failed?

@leofang
Copy link
Member

leofang commented Jan 16, 2026

I don't have direct access to the CI logs from commit 1fb2e77. Could you please specify which specific tests failed? I skipped these 11 tests:

@copilot Only 5 tests failed:

=========================== short test summary info ============================
FAILED cudapy/test_array_reductions.py::TestArrayReductions::test_mean_basic - cuda.bindings.nvjitlink.nvJitLinkError: ERROR_PTX_COMPILE (4)
Linker error log: ptxas fatal   : (C7907) Internal compiler error.
ptxas fatal   : Ptx assembly aborted due to errors
ERROR NVJITLINK_ERROR_PTX_COMPILE: JIT the PTX (ltoPtx)

FAILED cudapy/test_array_reductions.py::TestArrayReductions::test_nanmean_basic - cuda.bindings.nvjitlink.nvJitLinkError: ERROR_PTX_COMPILE (4)
Linker error log: ptxas fatal   : (C7907) Internal compiler error.
ptxas fatal   : Ptx assembly aborted due to errors
ERROR NVJITLINK_ERROR_PTX_COMPILE: JIT the PTX (ltoPtx)

FAILED cudapy/test_array_reductions.py::TestArrayReductions::test_nanprod_basic - cuda.bindings.nvjitlink.nvJitLinkError: ERROR_PTX_COMPILE (4)
Linker error log: ptxas fatal   : (C7907) Internal compiler error.
ptxas fatal   : Ptx assembly aborted due to errors
ERROR NVJITLINK_ERROR_PTX_COMPILE: JIT the PTX (ltoPtx)

FAILED cudapy/test_array_reductions.py::TestArrayReductions::test_nansum_basic - cuda.bindings.nvjitlink.nvJitLinkError: ERROR_PTX_COMPILE (4)
Linker error log: ptxas fatal   : (C7907) Internal compiler error.
ptxas fatal   : Ptx assembly aborted due to errors
ERROR NVJITLINK_ERROR_PTX_COMPILE: JIT the PTX (ltoPtx)

FAILED cudapy/test_array_reductions.py::TestArrayReductions::test_sum_basic - cuda.bindings.nvjitlink.nvJitLinkError: ERROR_PTX_COMPILE (4)
Linker error log: ptxas fatal   : (C7907) Internal compiler error.
ptxas fatal   : Ptx assembly aborted due to errors
ERROR NVJITLINK_ERROR_PTX_COMPILE: JIT the PTX (ltoPtx)

====== 5 failed, 1909 passed, 47 skipped, 14 xfailed in 333.45s (0:05:33) ======

Copy link
Contributor Author

Copilot AI commented Jan 16, 2026

I don't have direct access to the CI logs from commit 1fb2e77. Could you please specify which specific tests f...

Fixed in commit 56362b4. Removed the skip decorator from the 6 tests that didn't fail (test_var_basic, test_std_basic, test_min_basic, test_max_basic, test_nanmin_basic, test_nanmax_basic). Now only the 5 tests that actually failed in the CI logs are skipped: test_sum_basic, test_mean_basic, test_nansum_basic, test_nanmean_basic, and test_nanprod_basic.

@leofang
Copy link
Member

leofang commented Jan 16, 2026

/ok to test 56362b4

@leofang
Copy link
Member

leofang commented Jan 16, 2026

Copy link
Contributor Author

Copilot AI commented Jan 16, 2026

@copilot please fix the pre-commit error, see log at
https://github.com/NVIDIA/numba-cuda/actions/runs/21057282189/job/60555898305?pr=705

Fixed in commit 239d825. The issue was that ruff-format needed to reformat the skip decorator lines because they exceeded the line length limit. The decorator strings are now split across multiple lines to comply with formatting requirements. All pre-commit checks now pass.

@leofang
Copy link
Member

leofang commented Jan 16, 2026

/ok to test 239d825

@leofang leofang force-pushed the copilot/sync-ci-directory-for-cuda-13-1 branch from 239d825 to 76fa015 Compare January 16, 2026 06:36
@leofang
Copy link
Member

leofang commented Jan 16, 2026

/ok to test 76fa015

@leofang
Copy link
Member

leofang commented Jan 16, 2026

/ok to test 7d3f80a

@leofang leofang added the 3 - Ready for Review Ready for review by team label Jan 16, 2026
Copy link
Contributor

@rparolin rparolin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm.

Request: Can you update PR description with why nansum is failing on sm120? Thank you!

@leofang leofang merged commit 57b675b into main Jan 16, 2026
106 checks passed
@leofang leofang deleted the copilot/sync-ci-directory-for-cuda-13-1 branch January 16, 2026 21:01
@leofang
Copy link
Member

leofang commented Jan 16, 2026

Request: Can you update PR description with why nansum is failing on sm120? Thank you!

#725

gmarkall added a commit to gmarkall/numba-cuda that referenced this pull request Jan 27, 2026
- Add Python 3.14 to the wheel publishing matrix (NVIDIA#750)
- feat: swap out internal device array usage with `StridedMemoryView` (NVIDIA#703)
- Fix max block size computation in `forall` (NVIDIA#744)
- Fix prologue debug line info pointing to decorator instead of def line (NVIDIA#746)
- Fix kernel return type in DISubroutineType debug metadata (NVIDIA#745)
- Fix missing line info in Jupyter notebooks (NVIDIA#742)
- Fix: Pass correct flags to linker when debugging in the presence of LTOIR code (NVIDIA#698)
- chore(deps): add cuda-pathfinder to pixi deps (NVIDIA#741)
- fix: enable flake8-bugbear lints and fix found problems (NVIDIA#708)
- fix: Fix race condition in CUDA Simulator (NVIDIA#690)
- ci: run tests in parallel (NVIDIA#740)
- feat: users can pass `shared_memory_carveout` to @cuda.jit (NVIDIA#642)
- Fix compatibility with NumPy 2.4: np.trapz and np.in1d removed (NVIDIA#739)
- Pass the -numba-debug flag to libnvvm (NVIDIA#681)
- ci: remove rapids containers from conda ci (NVIDIA#737)
- Use `pathfinder` for dynamic libraries (NVIDIA#308)
- CI: Add CUDA 13.1 testing support (NVIDIA#705)
- Adding `pixi run test` and `pixi run test-par` support (NVIDIA#724)
- Disable per-PR nvmath tests + follow same test practice (NVIDIA#723)
- chore(deps): regenerate pixi lockfile (NVIDIA#722)
- Fix DISubprogram line number to point to function definition line (NVIDIA#695)
- revert: chore(dev): build pixi using rattler (NVIDIA#713) (NVIDIA#719)
- [feat] Initial version of the Numba CUDA GDB pretty-printer (NVIDIA#692)
- chore(dev): build pixi using rattler (NVIDIA#713)
- build(deps): bump the actions-monthly group across 1 directory with 8 updates (NVIDIA#704)
@gmarkall gmarkall mentioned this pull request Jan 27, 2026
kkraus14 pushed a commit that referenced this pull request Jan 28, 2026
- Add Python 3.14 to the wheel publishing matrix (#750)
- feat: swap out internal device array usage with `StridedMemoryView`
(#703)
- Fix max block size computation in `forall` (#744)
- Fix prologue debug line info pointing to decorator instead of def line
(#746)
- Fix kernel return type in DISubroutineType debug metadata (#745)
- Fix missing line info in Jupyter notebooks (#742)
- Fix: Pass correct flags to linker when debugging in the presence of
LTOIR code (#698)
- chore(deps): add cuda-pathfinder to pixi deps (#741)
- fix: enable flake8-bugbear lints and fix found problems (#708)
- fix: Fix race condition in CUDA Simulator (#690)
- ci: run tests in parallel (#740)
- feat: users can pass `shared_memory_carveout` to @cuda.jit (#642)
- Fix compatibility with NumPy 2.4: np.trapz and np.in1d removed (#739)
- Pass the -numba-debug flag to libnvvm (#681)
- ci: remove rapids containers from conda ci (#737)
- Use `pathfinder` for dynamic libraries (#308)
- CI: Add CUDA 13.1 testing support (#705)
- Adding `pixi run test` and `pixi run test-par` support (#724)
- Disable per-PR nvmath tests + follow same test practice (#723)
- chore(deps): regenerate pixi lockfile (#722)
- Fix DISubprogram line number to point to function definition line
(#695)
- revert: chore(dev): build pixi using rattler (#713) (#719)
- [feat] Initial version of the Numba CUDA GDB pretty-printer (#692)
- chore(dev): build pixi using rattler (#713)
- build(deps): bump the actions-monthly group across 1 directory with 8
updates (#704)

<!--

Thank you for contributing to numba-cuda :)

Here are some guidelines to help the review process go smoothly.

1. Please write a description in this text box of the changes that are
being
   made.

2. Please ensure that you have written units tests for the changes
made/features
   added.

3. If you are closing an issue please use one of the automatic closing
words as
noted here:
https://help.github.com/articles/closing-issues-using-keywords/

4. If your pull request is not ready for review but you want to make use
of the
continuous integration testing facilities please label it with `[WIP]`.

5. If your pull request is ready to be reviewed without requiring
additional
work on top of it, then remove the `[WIP]` label (if present) and
replace
it with `[REVIEW]`. If assistance is required to complete the
functionality,
for example when the C/C++ code of a feature is complete but Python
bindings
are still required, then add the label `[HELP-REQ]` so that others can
triage
and assist. The additional changes then can be implemented on top of the
same PR. If the assistance is done by members of the rapidsAI team, then
no
additional actions are required by the creator of the original PR for
this,
otherwise the original author of the PR needs to give permission to the
person(s) assisting to commit to their personal fork of the project. If
that
doesn't happen then a new PR based on the code of the original PR can be
opened by the person assisting, which then will be the PR that will be
   merged.

6. Once all work has been done and review has taken place please do not
add
features or make changes out of the scope of those requested by the
reviewer
(doing this just add delays as already reviewed code ends up having to
be
re-reviewed/it is hard to tell what is new etc!). Further, please do not
rebase your branch on main/force push/rewrite history, doing any of
these
   causes the context of any comments made by reviewers to be lost. If
   conflicts occur against main they should be resolved by merging main
   into the branch used for making the pull request.

Many thanks in advance for your cooperation!

-->
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

3 - Ready for Review Ready for review by team

Projects

None yet

Development

Successfully merging this pull request may close these issues.

CI: Test CUDA 13.1

4 participants