Skip to content

UPSTREAM PR #19117: server: print actual model name in 'model not found" error#1130

Open
loci-dev wants to merge 1 commit intomainfrom
loci/pr-19117-teto-print-model-name-when-not-found
Open

UPSTREAM PR #19117: server: print actual model name in 'model not found" error#1130
loci-dev wants to merge 1 commit intomainfrom
loci/pr-19117-teto-print-model-name-when-not-found

Conversation

@loci-dev
Copy link

@loci-dev loci-dev commented Feb 2, 2026

Note

Source pull request: ggml-org/llama.cpp#19117

Experimenting with AI, my environment gets messy fast and it's not always easy to know what model my software is trying to load. This helps with troubleshooting.

Before:

> Who are you ?


Error: {
  code = 400,
  message = "model not found",
  type = "invalid_request_error"
}

after

> Who are you ?


Error: {
  code = 400,
  message = "model 'toto' is not found",
  type = "invalid_request_error"
}

NB: I couldn't find a target to run linting (usually make lint / or make format) so tried to run the CI locally as explained in contributing.md with

GG_BUILD_CUDA=1 bash ./ci/run.sh ./tmp/results ./tmp/mnt
Warning: Using fallback CUDA architectures
./ci/run.sh: ligne 651: python3 : commande introuvable

I am developing from within the repo flake.nix so I expected all devtools present (I noticed the flake didn't provide ccache either, not sure if that's a conscious decision or not). I can submit a patch to add python3.

Make sure to read the contributing guidelines before submitting a PR

Experimenting with AI, my environment gets messy fast and it's not
always easy to know what model my software is trying to load. This helps
with troubleshooting.

before:

Error: {
  code = 400,
  message = "model not found",
  type = "invalid_request_error"
}

After:

Error: {
  code = 400,
  message = "model 'toto' not found",
  type = "invalid_request_error"
}
@loci-review
Copy link

loci-review bot commented Feb 2, 2026

No meaningful performance changes were detected across 115327 analyzed functions in the following binaries: build.bin.libllama.so, build.bin.llama-tts, build.bin.llama-cvector-generator, build.bin.libmtmd.so, build.bin.llama-tokenize, build.bin.llama-quantize, build.bin.llama-qwen2vl-cli, build.bin.libggml.so, build.bin.libggml-cpu.so, build.bin.libggml-base.so, build.bin.llama-bench, build.bin.llama-gguf-split, build.bin.llama-llava-cli, build.bin.llama-minicpmv-cli, build.bin.llama-gemma3-cli.

🔎 Full breakdown: Loci Inspector.
💬 Questions? Tag @loci-dev.

@loci-dev loci-dev force-pushed the main branch 27 times, most recently from 99b11e9 to cd152fa Compare February 3, 2026 10:20
@loci-dev loci-dev force-pushed the main branch 10 times, most recently from 2cecc98 to a92fe2a Compare February 26, 2026 02:16
@loci-dev loci-dev force-pushed the main branch 8 times, most recently from 9f4f332 to 4298c74 Compare March 6, 2026 02:17
@loci-dev loci-dev force-pushed the main branch 9 times, most recently from 56aaa36 to 21147c2 Compare March 13, 2026 02:17
@loci-dev loci-dev force-pushed the main branch 3 times, most recently from f2637dc to 910a8a6 Compare March 16, 2026 02:18
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants