You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
also, for the previous version, for the model_type keyword, there are qwen2_14b-instruct and qwen2_14b to distinct the template between chat and base, but in this 3.x version, we only use the qwen2_5
The text was updated successfully, but these errors were encountered:
Describe the bug
The eos token does not work after sft the qwen2.5-14b base model.
Only if I add the stop token, the finish_reason can be stop.
the training script is like below, I use the base model to do sft training:
the sft data is like below
![Image](https://private-user-images.githubusercontent.com/3861285/413186768-708b2b14-d179-4f79-9a7e-ec1cad4fa9bc.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk2MjQ1MjksIm5iZiI6MTczOTYyNDIyOSwicGF0aCI6Ii8zODYxMjg1LzQxMzE4Njc2OC03MDhiMmIxNC1kMTc5LTRmNzktOWE3ZS1lYzFjYWQ0ZmE5YmMucG5nP1gtQW16LUFsZ29yaXRobT1BV1M0LUhNQUMtU0hBMjU2JlgtQW16LUNyZWRlbnRpYWw9QUtJQVZDT0RZTFNBNTNQUUs0WkElMkYyMDI1MDIxNSUyRnVzLWVhc3QtMSUyRnMzJTJGYXdzNF9yZXF1ZXN0JlgtQW16LURhdGU9MjAyNTAyMTVUMTI1NzA5WiZYLUFtei1FeHBpcmVzPTMwMCZYLUFtei1TaWduYXR1cmU9NDY0N2I4NWFiMzVmMDAwZTE4ODYwMmY3NmRiMGQxOGMxMWEzYjc2MDJlOWEwYzZlMjc4NTQyOGUzNzQ0ZDMxZSZYLUFtei1TaWduZWRIZWFkZXJzPWhvc3QifQ.arjJf250KwYtLtW8Z46SLNqeXKuV1D0t0AckHR5Tw0s)
Your hardware and system info
8*A800 80G
ms-swift 3.0.2.post1
torch 2.4.0
python 3.9
transformers 4.45.2
Additional context
check the config file of the saving model, the eos_token_id in generation_config.json is not correct, as well as the config.json.
also, for the previous version, for the model_type keyword, there are qwen2_14b-instruct and qwen2_14b to distinct the template between chat and base, but in this 3.x version, we only use the qwen2_5
The text was updated successfully, but these errors were encountered: