Skip to content

Conversation

hynky1999
Copy link
Collaborator

@hynky1999 hynky1999 commented Feb 6, 2025

Fixes #532

@hynky1999 hynky1999 requested review from lewtun and NathanHB February 6, 2025 12:28
@HuggingFaceDocBuilderDev
Copy link
Collaborator

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

# note: this has changed on 0.3.3, and it only works now if num_gpus are set.
# but then tensor_parallel breaks
@ray.remote
# Hynek: With the newest vllm, it actually breaks when tensor_parallel_size == 1 and num_gpus not set,
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do we need to fix vllm version?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah yes, it should be lower-bounded from v0.7.0

}
if int(config.data_parallel_size) > 1:
self.model_args["worker_use_ray"] = True
self.model_args["distributed_executor_backend"] = "ray"
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Will this work with both versions of vllm?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It will break for versions less than v0.7.0. A solution would be to have an if/else statement that checks the vllm version and falls back to the old logic. Personally I would just lower bound vllm since it's gonna be a pain to keep track of all the changes over time

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Would work for me

Copy link
Member

@lewtun lewtun left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Great fix, LGTM!

@clefourrier
Copy link
Member

Don't merge before fixing vllm import ^^

@hynky1999
Copy link
Collaborator Author

should be good now :)

@lewtun
Copy link
Member

lewtun commented Feb 6, 2025

@hynky1999 hynky1999 merged commit 86f6225 into main Feb 6, 2025
4 checks passed
hynky1999 added a commit that referenced this pull request May 22, 2025
* make bleur lazy

* make tokenizer lazy too

* fix_ray

* fix tensor_paralel > 1

* remove debug statements

* bump vllm

---------

Co-authored-by: Hynek Kydlicek <[email protected]>
NathanHB pushed a commit that referenced this pull request Sep 19, 2025
* make bleur lazy

* make tokenizer lazy too

* fix_ray

* fix tensor_paralel > 1

* remove debug statements

* bump vllm

---------

Co-authored-by: Hynek Kydlicek <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

evaluation error: TypeError: EngineArgs.__init__() got an unexpected keyword argument 'worker_use_ray'
4 participants