Skip to content

Align use of vllm_max_model_length in RLOOTrainer#4702

Merged
albertvillanova merged 2 commits intohuggingface:mainfrom
albertvillanova:align-vllm-max-model-length
Dec 16, 2025
Merged

Align use of vllm_max_model_length in RLOOTrainer#4702
albertvillanova merged 2 commits intohuggingface:mainfrom
albertvillanova:align-vllm-max-model-length

Conversation

@albertvillanova
Copy link
Member

Align use of vllm_max_model_length in RLOOTrainer.

Align it with GRPOTrainer, introduced in PR:

This alignment is required to make simpler this PR:

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Control the GPU memory utilization for vLLM. This setting only applies when `vllm_mode` is set to
`"colocate"`. If you are using `vllm_mode="server"`, this parameter must be passed separately when
launching the vLLM server via the `--vllm_gpu_memory_utilization` flag.
vllm_max_model_length (`int`, *optional*, defaults to `None`):
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
vllm_max_model_length (`int`, *optional*, defaults to `None`):
vllm_max_model_length (`int`, *optional*):

can you possibly apply the same in grpo config?

Copy link
Member

@qgallouedec qgallouedec left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm!

@albertvillanova albertvillanova merged commit 61c9921 into huggingface:main Dec 16, 2025
8 of 9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants