Skip to content

UPSTREAM PR #18413: ggml-cuda: use CMAKE_CUDA_ARCHITECTURES if set when GGML_NATIVE=ON#722

Open
loci-dev wants to merge 1 commit intomainfrom
upstream-PR18413-branch_QDelta-master
Open

UPSTREAM PR #18413: ggml-cuda: use CMAKE_CUDA_ARCHITECTURES if set when GGML_NATIVE=ON#722
loci-dev wants to merge 1 commit intomainfrom
upstream-PR18413-branch_QDelta-master

Conversation

@loci-dev
Copy link
Copy Markdown

Mirrored from ggml-org/llama.cpp#18413

Fixed compilation error when building with GGML_NATIVE=ON and CUDA Toolkit but without a GPU (e.g. docker build) where CMAKE_CUDA_ARCHITECTURES_NATIVE will be No CUDA Devices found., causing nvcc error Unsupported gpu architecture 'compute_No CUDA Devices found.'.

Also if CMAKE_CUDA_ARCHITECTURES is explicitly set by the user, maybe it's better to use it even when GGML_NATIVE=ON. Which is also consistent with the behavior before #18361.

Additionally moves the logging of CMAKE_CUDA_ARCHITECTURES after the replacement.

Possibly related to #18398.

@loci-review
Copy link
Copy Markdown

loci-review bot commented Dec 27, 2025

Explore the complete analysis inside the Version Insights

Perfect! I've generated the summary report for your project. Here's what the analysis shows:

Summary Report for llama.cpp PR #722

Key Findings:

  • No significant performance impact detected
  • All modified functions show less than 2% change in both response time and throughput
  • The pull request is performance-neutral and safe to merge from a performance perspective

Project Information:

The analysis indicates that the changes in this PR maintain the existing performance profile of the application without introducing any performance regressions.

@loci-dev loci-dev force-pushed the main branch 27 times, most recently from bb623bb to 7f36157 Compare December 31, 2025 01:37
@loci-dev loci-dev force-pushed the main branch 30 times, most recently from 118039a to cd122e2 Compare January 6, 2026 07:13
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants