-
Notifications
You must be signed in to change notification settings - Fork 408
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
HumanEval doesn't actually runs test #1102
Comments
Up, the problem is still here An example of generated 'function' here:
And then example of
This code obviously won't do anything. Moreover, current implementation of: deepeval/deepeval/benchmarks/human_eval/human_eval.py Lines 93 to 102 in f8d5f8f
|
Stumbled upon the same issue, HumanEval isn't, in fact, supported |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
HumanEval.evaluate checks only if generated sample is correct python.
It doesn't actually run tests like the original project does.
deepeval/deepeval/benchmarks/human_eval/human_eval.py
Lines 96 to 97 in f8d5f8f
The text was updated successfully, but these errors were encountered: