Skip to content

Conversation

@Stillerman
Copy link
Contributor

@Stillerman Stillerman commented Jul 20, 2024

Adding SmolLM pre-tokenizer support for SmolLM models.

Added tokenizer type for SmolLM-135M in convert-hf-to-gguf-update.py
Added the chkhsh for SmolLM-135M in convert-hf-to-gguf.py
Added LLAMA_VOCAB_PRE_TYPE_SMOLLM enum to llama.h

Ran ./tests/test-tokenizer-0.sh smollm ./models/ggml-vocab-smollm.gguf and Tests passed.

Thank you @m18coppola for #8579

@loubnabnl @anton-l Does src/llama.cpp look right? Any special settings needed in the tokenizer? The .gguf I created for SmolLM-135M seemed to inference well.

@github-actions github-actions bot added the python python script changes label Jul 20, 2024
Stillerman and others added 2 commits July 21, 2024 02:48
@ngxson ngxson linked an issue Jul 21, 2024 that may be closed by this pull request
4 tasks
Copy link
Collaborator

@Vaibhavs10 Vaibhavs10 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Tested this! Seems to work well for me too!

@eliebak
Copy link

eliebak commented Jul 22, 2024

Seems to work well for me too @Stillerman, thanks a lot for adding support to our model! 🤗

@spergware
Copy link

The 1.7B model prints out nonsense, isn't that supported? Asking for clarification.

arthw pushed a commit to arthw/llama.cpp that referenced this pull request Jul 27, 2024
* Adding SmolLM Pre Tokenizer

* Update convert_hf_to_gguf_update.py

Co-authored-by: compilade <[email protected]>

* Update src/llama.cpp

Co-authored-by: compilade <[email protected]>

* handle regex

* removed .inp and out .out ggufs

---------

Co-authored-by: compilade <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

python python script changes

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Support for SmolLM

6 participants