Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deps(whisper.cpp): update, fix cublas build #1846

Merged
merged 1 commit into from
Mar 18, 2024
Merged

Conversation

mudler
Copy link
Owner

@mudler mudler commented Mar 16, 2024

Description

This PR fixes #1733

Notes for Reviewers

See also: https://gitlab.kitware.com/cmake/cmake/-/issues/25536 - looks a cmake issue (?)
upstream: ggerganov/whisper.cpp#1553

PR (whisper.cpp): ggerganov/whisper.cpp#1973

Signed commits

  • Yes, I signed my commits.

Copy link

netlify bot commented Mar 16, 2024

Deploy Preview for localai canceled.

Name Link
🔨 Latest commit 9ae3e67
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/65f82d5a440bde00086dd17f

@mudler
Copy link
Owner Author

mudler commented Mar 18, 2024

opening a PR upstream with the fix too: ggerganov/whisper.cpp#1973

for reference, this is solved by adding -L$(CUDA_PATH)/stubs -lcuda to the whisper LDFLAGS and in CGO_LDFLAGS when building the golang binary with libwhisper.a (not the .so).

I've traced it back by trying to find the symbols that it complains about:

nm -D /usr/local/cuda/targets/x86_64-linux/lib/*.so   | grep cuMem
<no results>

And instead present in the stubs directory:

root@76d08ab315dc:/build# nm -D /usr/local/cuda/targets/x86_64-linux/lib/stubs/*.so   | grep cuMem                                                                                                                                           
0000000000008000 T cuMemAddressFree         
0000000000007ff0 T cuMemAddressReserve    
00000000000081d0 T cuMemAdvise                                                                                        
0000000000008f30 T cuMemAlloc
0000000000009710 T cuMemAllocAsync
00000000000080e0 T cuMemAllocAsync_ptsz
0000000000009720 T cuMemAllocFromPoolAsync
0000000000008160 T cuMemAllocFromPoolAsync_ptsz
0000000000008f70 T cuMemAllocHost
0000000000007bf0 T cuMemAllocHost_v2
0000000000007c40 T cuMemAllocManaged
0000000000008f40 T cuMemAllocPitch
... (and many others)

See also: NVIDIA/nvidia-docker#508

@mudler mudler force-pushed the whisper_update_cublas branch 5 times, most recently from ad9d268 to 1c8b469 Compare March 18, 2024 11:43
@mudler mudler merged commit b202bfa into master Mar 18, 2024
29 checks passed
@mudler mudler deleted the whisper_update_cublas branch March 18, 2024 14:56
@mudler mudler added the enhancement New feature or request label Mar 18, 2024
truecharts-admin referenced this pull request in truecharts/public Mar 19, 2024
…0.1 by renovate (#19487)

This PR contains the following updates:

| Package | Update | Change |
|---|---|---|
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0-cublas-cuda11-ffmpeg-core` ->
`v2.10.1-cublas-cuda11-ffmpeg-core` |
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0-cublas-cuda11-core` -> `v2.10.1-cublas-cuda11-core` |
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0-cublas-cuda12-ffmpeg-core` ->
`v2.10.1-cublas-cuda12-ffmpeg-core` |
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0-cublas-cuda12-core` -> `v2.10.1-cublas-cuda12-core` |
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0-ffmpeg-core` -> `v2.10.1-ffmpeg-core` |
| [docker.io/localai/localai](https://github.com/mudler/LocalAI) |
patch | `v2.10.0` -> `v2.10.1` |

---

> [!WARNING]
> Some dependencies could not be looked up. Check the Dependency
Dashboard for more information.

---

### Release Notes

<details>
<summary>mudler/LocalAI (docker.io/localai/localai)</summary>

###
[`v2.10.1`](https://github.com/mudler/LocalAI/releases/tag/v2.10.1)

[Compare
Source](https://github.com/mudler/LocalAI/compare/v2.10.0...v2.10.1)

<!-- Release notes generated using configuration in .github/release.yml
at master -->

##### What's Changed

##### Bug fixes 🐛

- fix(llama.cpp): fix eos without cache by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1852](https://github.com/mudler/LocalAI/pull/1852)
- fix(config): default to debug=false if not set by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1853](https://github.com/mudler/LocalAI/pull/1853)
- fix(config-watcher): start only if config-directory exists by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1854](https://github.com/mudler/LocalAI/pull/1854)

##### Exciting New Features 🎉

- deps(whisper.cpp): update, fix cublas build by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1846](https://github.com/mudler/LocalAI/pull/1846)

##### Other Changes

- fixes [#&#8203;1051](https://github.com/mudler/LocalAI/issues/1051):
handle openai presence and request penalty parameters by
[@&#8203;blob42](https://github.com/blob42) in
[https://github.com/mudler/LocalAI/pull/1817](https://github.com/mudler/LocalAI/pull/1817)
- fix(make): allow to parallelize jobs by
[@&#8203;cryptk](https://github.com/cryptk) in
[https://github.com/mudler/LocalAI/pull/1845](https://github.com/mudler/LocalAI/pull/1845)
- fix(go-llama): use llama-cpp as default by
[@&#8203;mudler](https://github.com/mudler) in
[https://github.com/mudler/LocalAI/pull/1849](https://github.com/mudler/LocalAI/pull/1849)
- ⬆️ Update docs version mudler/LocalAI by
[@&#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1847](https://github.com/mudler/LocalAI/pull/1847)
- ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://github.com/localai-bot) in
[https://github.com/mudler/LocalAI/pull/1848](https://github.com/mudler/LocalAI/pull/1848)
- test/fix: OSX Test Repair by
[@&#8203;dave-gray101](https://github.com/dave-gray101) in
[https://github.com/mudler/LocalAI/pull/1843](https://github.com/mudler/LocalAI/pull/1843)

**Full Changelog**:
mudler/LocalAI@v2.10.0...v2.10.1

</details>

---

### Configuration

📅 **Schedule**: Branch creation - At any time (no schedule defined),
Automerge - At any time (no schedule defined).

🚦 **Automerge**: Enabled.

♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the
rebase/retry checkbox.

🔕 **Ignore**: Close this PR and you won't be reminded about these
updates again.

---

- [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check
this box

---

This PR has been generated by [Renovate
Bot](https://github.com/renovatebot/renovate).

<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy4yNTYuMCIsInVwZGF0ZWRJblZlciI6IjM3LjI1Ni4wIiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIn0=-->
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
1 participant