Skip to content

Conversation

alozowski
Copy link
Collaborator

  • Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation.
  • Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification.
  • Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.

@alozowski alozowski requested a review from NathanHB March 28, 2025 00:36
@HuggingFaceDocBuilderDev
Copy link
Collaborator

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.


case _:
return lambda x: x
raise ValueError(f"Unsupported backend: {self.backend}")
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@NathanHB were you using this case for some specific use cases?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oh it was because I wanted the details to be saved even with a failure in the judge so that we don't have to rerun everything.
I forgot to add a warning that's my bad :/

@clefourrier
Copy link
Member

Cool PR! A couple nits, and I'd like @NathanHB to take a look since he knows the llm judge section better than I do, but very useful!

Copy link
Member

@clefourrier clefourrier left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for addressing the comments, lgtm.

@alozowski alozowski merged commit e611a91 into main Mar 31, 2025
4 checks passed
alozowski added a commit that referenced this pull request Mar 31, 2025
hynky1999 pushed a commit that referenced this pull request May 22, 2025
- Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation.
- Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification.
- Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.
NathanHB pushed a commit that referenced this pull request Sep 19, 2025
- Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation.
- Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification.
- Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants