Skip to content

Commit

Permalink
fix: Fix typos in evaluation example metric prompt templates.
Browse files Browse the repository at this point in the history
PiperOrigin-RevId: 676917898
  • Loading branch information
jsondai authored and copybara-github committed Sep 20, 2024
1 parent 2b84142 commit 5f4d586
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions vertexai/evaluation/metrics/_default_templates.py
Original file line number Diff line number Diff line change
Expand Up @@ -390,7 +390,7 @@
## Evaluation Steps
STEP 1: Analyze Response A based on the instruction following criteria: Determine how well Response A fulfills the requirements outlined in the instructions and provide assessment according to the criterion.
STEP 2: Analyze Response B based on the instruction following criteria: Determine how well Response A fulfills the requirements outlined in the instructions and provide assessment according to the criterion.
STEP 2: Analyze Response B based on the instruction following criteria: Determine how well Response B fulfills the requirements outlined in the instructions and provide assessment according to the criterion.
STEP 3: Compare the overall performance of Response A and Response B based on your analyses and assessment.
STEP 4: Output your preference of "A", "SAME" or "B" to the pairwise_choice field according to the Rating Rubric.
STEP 5: Output your assessment reasoning in the explanation field.
Expand Down Expand Up @@ -900,7 +900,7 @@
## Evaluation Steps
STEP 1: Analyze Response A based on the question answering quality criteria: Determine how well Response A fulfills the user requirements, is grounded in the context, is complete and fluent, and provides assessment according to the criterion.
STEP 2: Analyze Response B based on the question answering quality criteria: Determine how well Response A fulfills the user requirements, is grounded in the context, is complete and fluent, and provides assessment according to the criterion.
STEP 2: Analyze Response B based on the question answering quality criteria: Determine how well Response B fulfills the user requirements, is grounded in the context, is complete and fluent, and provides assessment according to the criterion.
STEP 3: Compare the overall performance of Response A and Response B based on your analyses and assessment.
STEP 4: Output your preference of "A", "SAME" or "B" to the pairwise_choice field according to the Rating Rubric.
STEP 5: Output your assessment reasoning in the explanation field.
Expand Down

0 comments on commit 5f4d586

Please sign in to comment.