You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
Currently, CounterfactualGenerator and AutoEval return a dictionary containing only metric values. In contrast, ToxicityMetrics and StereotypeMetrics give users the option to return response-level scores for a more detailed investigation into the assessment results. The request here is to give users the option to return analogous, response-level scores for CounterfactualGenerator, and return response-level toxicity, stereotype, and counterfactual scores with AutoEval.
Describe the solution you'd like
Give users option to return response-level scores with CounterfactualGenerator and AutoEval in returned dictionary. For consistency with ToxicityMetrics and StereotypeMetrics, this should be achieved by adding a new boolean argument called return_data to CounterfactualGenerator.evaluate and AutoEval.evaluate. The dictionary format returned should be consistent across these classes: a "metrics" key with metric values, and a "data" key with responses and response-level scores.
Additional context
Along with counterfactual.py, auto.py, demo notebooks, unit tests, and readme should also be updated.
The text was updated successfully, but these errors were encountered:
Is your feature request related to a problem? Please describe.
Currently,
CounterfactualGenerator
andAutoEval
return a dictionary containing only metric values. In contrast,ToxicityMetrics
andStereotypeMetrics
give users the option to return response-level scores for a more detailed investigation into the assessment results. The request here is to give users the option to return analogous, response-level scores forCounterfactualGenerator
, and return response-level toxicity, stereotype, and counterfactual scores withAutoEval
.Describe the solution you'd like
Give users option to return response-level scores with
CounterfactualGenerator
andAutoEval
in returned dictionary. For consistency withToxicityMetrics
andStereotypeMetrics
, this should be achieved by adding a new boolean argument calledreturn_data
toCounterfactualGenerator.evaluate
andAutoEval.evaluate
. The dictionary format returned should be consistent across these classes: a "metrics" key with metric values, and a "data" key with responses and response-level scores.Additional context
Along with counterfactual.py, auto.py, demo notebooks, unit tests, and readme should also be updated.
The text was updated successfully, but these errors were encountered: