llama-server: Support --verbose-prompt#19752
Conversation
llama-server has the option `--verbose-prompt` but was ineffective. This commit recovers the logger for full prompt input (gated behind the non-default option `--verbose-prompt`).
aa0925f to
da2e13c
Compare
|
Use slots debug instead |
|
Okay, undocumented slot debugging feature...? How to use it?Invoking with Slots Debug enabledLLAMA_SERVER_SLOTS_DEBUG=1 llama-server --port 11434 [ARGS...]How to inspect the results?After the process finishes, inspect # curl --silent 'http://localhost:11434/slots' | jq
[
...,
{
"id": 7,
"n_ctx": 2048,
"speculative": false,
"is_processing": false,
"id_task": 0,
...,
"prompt": "<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n<|im_start|>user\nHello!<|im_end|>\n<|im_start|>assistant\n",
"generated": "Hello! How can I assist you today?"
}
]You may need to specify the model name on the router mode. curl --silent 'http://localhost:11434/slots?model=Qwen%2FQwen2.5-Coder-1.5B-Instruct' | jqAlternatives@ngxson Then, I came up with one question. What's the point leaving What do you think? |
|
@a4lg It was undocumented because it was mean to be used for testing only. But feel free to add it to a new section at the end of server docs.
Yes, please, that would be helpful. You can push it with the same PR I suggested above |
llama-server has the option
--verbose-promptbut was ineffective.This commit recovers the logger for full prompt input (gated behind the non-default option
--verbose-prompt) which was originally commented out.