Skip to content

[rollout] fix: compute max_tokens based on response_length with promp…

e785d6e
Select commit
Loading
Failed to load commit list.
Closed

[vllm, fully_async] fix: clamp max_tokens to response_length instead of max_model_len - prompt_len in async vLLM rollout #5505

[rollout] fix: compute max_tokens based on response_length with promp…
e785d6e
Select commit
Loading
Failed to load commit list.

Workflow runs completed with no jobs