-
Notifications
You must be signed in to change notification settings - Fork 0
UPSTREAM PR #17005: CUDA: update ops.md #84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
UPSTREAM PR #17005: CUDA: update ops.md #84
Conversation
…unary_floor_round_ceil_trunc
|
Access the complete analysis in the LOCI Dashboard Performance Analysis SummaryOverviewAnalysis of version 971c7425 compared to baseline 523c96f3 reveals minimal performance variations within statistical noise thresholds. The highest observed changes were in non-core functions with negligible impact on inference performance. Key FindingsPerformance Metrics:
Core Function Impact: Inference Performance Impact: Power Consumption Analysis:
Flame Graph and CFG Analysis: GitHub Code Review: Conclusion: |
40efe8b to
3e9b10f
Compare
Mirrored from ggml-org/llama.cpp#17005
Forgot to update ops added by me.
@pwilkin @am17an
ref: ggml-org/llama.cpp#16917
ref: ggml-org/llama.cpp#15635