We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 8a476be commit 9b15621Copy full SHA for 9b15621
llama.h
@@ -413,7 +413,7 @@ extern "C" {
413
414
LLAMA_API bool llama_supports_mmap (void);
415
LLAMA_API bool llama_supports_mlock (void);
416
- LLAMA_API bool llama_supports_rpc (void);
+ LLAMA_API bool llama_supports_rpc (void); // TMP: https://github.com/ggerganov/llama.cpp/pull/7647#issuecomment-2140234367
417
LLAMA_API bool llama_supports_gpu_offload(void);
418
419
LLAMA_API const struct llama_model * llama_get_model(const struct llama_context * ctx);
0 commit comments