You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[https://nvbugs/5513423][fix] Correctly respect min_tokens in PyTorch workflow using TorchSampler
- Added `py_min_length` attribute to `LlmRequest` to store minimum length configuration.
- Implemented `_apply_min_length_penalty` method in `TorchSampler` to adjust logits based on minimum length requirements. (Mimics PenaltyLayer)
- Updated test case for `min_tokens` to reflect new maximum sequence length and output length constraints from the model
Signed-off-by: Stefan Niebler <[email protected]>
0 commit comments