Regarding the behavior of max_seq_length in SFTTrainer #2400
Labels
📚 documentation
Improvements or additions to documentation
👶 good first issue
Good for newcomers
🏋 SFT
Related to SFT
The SFTTrainer documentation states:
However, when looking at the actual code, the Tokenizer appears to have padding=False, and it does not seem to pad sequences to the max_seq_length value.
https://github.com/huggingface/trl/blob/main/trl/trainer/sft_trainer.py#L420
How does SFTTrainer ensure that sequences are padded to the max_seq_length value?
The text was updated successfully, but these errors were encountered: