Spaces:
Running
Running
Update common.py
Browse files
common.py
CHANGED
|
@@ -47,28 +47,28 @@ EVAL_DESCRIPTION = """
|
|
| 47 |
- Examples (Optional)
|
| 48 |
"""
|
| 49 |
|
| 50 |
-
DEFAULT_EVAL_PROMPT = """
|
| 51 |
|
| 52 |
Scoring Rubric:
|
| 53 |
-
Score 1: The
|
| 54 |
-
Score 2: The
|
| 55 |
-
Score 3: The
|
| 56 |
-
Score 4: The
|
| 57 |
-
Score 5: The
|
| 58 |
|
| 59 |
[User Query]: {{input}}
|
| 60 |
|
| 61 |
[AI Response]: {{response}}"""
|
| 62 |
|
| 63 |
# Split the eval prompt into editable and fixed parts
|
| 64 |
-
DEFAULT_EVAL_PROMPT_EDITABLE = """
|
| 65 |
|
| 66 |
Scoring Rubric:
|
| 67 |
-
Score 1: The
|
| 68 |
-
Score 2: The
|
| 69 |
-
Score 3: The
|
| 70 |
-
Score 4: The
|
| 71 |
-
Score 5: The
|
| 72 |
|
| 73 |
# Fixed suffix that will always be appended
|
| 74 |
FIXED_EVAL_SUFFIX = """
|
|
@@ -164,17 +164,17 @@ We’d love to hear your feedback! For general feature requests or to submit / s
|
|
| 164 |
|
| 165 |
|
| 166 |
# Default values for compatible mode
|
| 167 |
-
DEFAULT_EVAL_CRITERIA = """
|
| 168 |
|
| 169 |
-
DEFAULT_SCORE_1 = "The
|
| 170 |
|
| 171 |
-
DEFAULT_SCORE_2 = "The
|
| 172 |
|
| 173 |
-
DEFAULT_SCORE_3 = "The
|
| 174 |
|
| 175 |
-
DEFAULT_SCORE_4 = "The
|
| 176 |
|
| 177 |
-
DEFAULT_SCORE_5 = "The
|
| 178 |
|
| 179 |
#**What are the Evaluator Prompt Templates based on?**
|
| 180 |
|
|
|
|
| 47 |
- Examples (Optional)
|
| 48 |
"""
|
| 49 |
|
| 50 |
+
DEFAULT_EVAL_PROMPT = """Does the model provide relevant and useful responses to the user's needs or questions?
|
| 51 |
|
| 52 |
Scoring Rubric:
|
| 53 |
+
Score 1: The model's responses are irrelevant or unhelpful to the user's needs or queries.
|
| 54 |
+
Score 2: The model sometimes provides helpful information, but often fails to address the user's actual needs or questions.
|
| 55 |
+
Score 3: The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark.
|
| 56 |
+
Score 4: The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies.
|
| 57 |
+
Score 5: The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries.
|
| 58 |
|
| 59 |
[User Query]: {{input}}
|
| 60 |
|
| 61 |
[AI Response]: {{response}}"""
|
| 62 |
|
| 63 |
# Split the eval prompt into editable and fixed parts
|
| 64 |
+
DEFAULT_EVAL_PROMPT_EDITABLE = """Does the model provide relevant and useful responses to the user's needs or questions?
|
| 65 |
|
| 66 |
Scoring Rubric:
|
| 67 |
+
Score 1: The model's responses are irrelevant or unhelpful to the user's needs or queries.
|
| 68 |
+
Score 2: The model sometimes provides helpful information, but often fails to address the user's actual needs or questions.
|
| 69 |
+
Score 3: The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark.
|
| 70 |
+
Score 4: The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies.
|
| 71 |
+
Score 5: The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries."""
|
| 72 |
|
| 73 |
# Fixed suffix that will always be appended
|
| 74 |
FIXED_EVAL_SUFFIX = """
|
|
|
|
| 164 |
|
| 165 |
|
| 166 |
# Default values for compatible mode
|
| 167 |
+
DEFAULT_EVAL_CRITERIA = """Does the model provide relevant and useful responses to the user's needs or questions?"""
|
| 168 |
|
| 169 |
+
DEFAULT_SCORE_1 = "The model's responses are irrelevant or unhelpful to the user's needs or queries."
|
| 170 |
|
| 171 |
+
DEFAULT_SCORE_2 = "The model sometimes provides helpful information, but often fails to address the user's actual needs or questions."
|
| 172 |
|
| 173 |
+
DEFAULT_SCORE_3 = "The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark."
|
| 174 |
|
| 175 |
+
DEFAULT_SCORE_4 = "The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies."
|
| 176 |
|
| 177 |
+
DEFAULT_SCORE_5 = "The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries."
|
| 178 |
|
| 179 |
#**What are the Evaluator Prompt Templates based on?**
|
| 180 |
|