Skip to content

Conversation

@aldehir
Copy link

@aldehir aldehir commented Aug 31, 2025

Add support for nvidia/NVIDIA-Nemotron-Nano-9B-v2.

The template strips the last message if role == "assistant", causing it to fail the tool call checks. I added tool responses to make it happy, but I understand this somewhat overlaps with the tool call response check. I'm happy to change the implementation if there is a better alternative.

ref: ggml-org/llama.cpp#15676 (comment)

@pwilkin
Copy link

pwilkin commented Aug 31, 2025

One obvious solution would be to pad the tool call testing message with another assistant message at the end.

Edit: nvm, didn't look at the commits, I see you already did that :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants