-
Notifications
You must be signed in to change notification settings - Fork 361
Integrate huggingface_hub inference support #651
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
|
||
case _: | ||
return lambda x: x | ||
raise ValueError(f"Unsupported backend: {self.backend}") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@NathanHB were you using this case for some specific use cases?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oh it was because I wanted the details to be saved even with a failure in the judge so that we don't have to rerun everything.
I forgot to add a warning that's my bad :/
Cool PR! A couple nits, and I'd like @NathanHB to take a look since he knows the llm judge section better than I do, but very useful! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for addressing the comments, lgtm.
This reverts commit e611a91.
- Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation. - Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification. - Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.
- Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation. - Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification. - Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.
hf-inference
) as a new backend option for judge model evaluation.AsyncInferenceClient
only forhf-inference
, with support for retries, timeouts, and provider specification.JudgeLM
andJudgeLLM
classes to support new configuration parameters likehf_provider
,base_url
, andmax_tokens
.