Spaces:
Running
Running
Update common.py
Browse files
common.py
CHANGED
@@ -47,28 +47,28 @@ EVAL_DESCRIPTION = """
|
|
47 |
- Examples (Optional)
|
48 |
"""
|
49 |
|
50 |
-
DEFAULT_EVAL_PROMPT = """
|
51 |
|
52 |
Scoring Rubric:
|
53 |
-
Score 1: The
|
54 |
-
Score 2: The
|
55 |
-
Score 3: The
|
56 |
-
Score 4: The
|
57 |
-
Score 5: The
|
58 |
|
59 |
[User Query]: {{input}}
|
60 |
|
61 |
[AI Response]: {{response}}"""
|
62 |
|
63 |
# Split the eval prompt into editable and fixed parts
|
64 |
-
DEFAULT_EVAL_PROMPT_EDITABLE = """
|
65 |
|
66 |
Scoring Rubric:
|
67 |
-
Score 1: The
|
68 |
-
Score 2: The
|
69 |
-
Score 3: The
|
70 |
-
Score 4: The
|
71 |
-
Score 5: The
|
72 |
|
73 |
# Fixed suffix that will always be appended
|
74 |
FIXED_EVAL_SUFFIX = """
|
@@ -164,17 +164,17 @@ We’d love to hear your feedback! For general feature requests or to submit / s
|
|
164 |
|
165 |
|
166 |
# Default values for compatible mode
|
167 |
-
DEFAULT_EVAL_CRITERIA = """
|
168 |
|
169 |
-
DEFAULT_SCORE_1 = "The
|
170 |
|
171 |
-
DEFAULT_SCORE_2 = "The
|
172 |
|
173 |
-
DEFAULT_SCORE_3 = "The
|
174 |
|
175 |
-
DEFAULT_SCORE_4 = "The
|
176 |
|
177 |
-
DEFAULT_SCORE_5 = "The
|
178 |
|
179 |
#**What are the Evaluator Prompt Templates based on?**
|
180 |
|
|
|
47 |
- Examples (Optional)
|
48 |
"""
|
49 |
|
50 |
+
DEFAULT_EVAL_PROMPT = """Does the model provide relevant and useful responses to the user's needs or questions?
|
51 |
|
52 |
Scoring Rubric:
|
53 |
+
Score 1: The model's responses are irrelevant or unhelpful to the user's needs or queries.
|
54 |
+
Score 2: The model sometimes provides helpful information, but often fails to address the user's actual needs or questions.
|
55 |
+
Score 3: The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark.
|
56 |
+
Score 4: The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies.
|
57 |
+
Score 5: The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries.
|
58 |
|
59 |
[User Query]: {{input}}
|
60 |
|
61 |
[AI Response]: {{response}}"""
|
62 |
|
63 |
# Split the eval prompt into editable and fixed parts
|
64 |
+
DEFAULT_EVAL_PROMPT_EDITABLE = """Does the model provide relevant and useful responses to the user's needs or questions?
|
65 |
|
66 |
Scoring Rubric:
|
67 |
+
Score 1: The model's responses are irrelevant or unhelpful to the user's needs or queries.
|
68 |
+
Score 2: The model sometimes provides helpful information, but often fails to address the user's actual needs or questions.
|
69 |
+
Score 3: The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark.
|
70 |
+
Score 4: The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies.
|
71 |
+
Score 5: The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries."""
|
72 |
|
73 |
# Fixed suffix that will always be appended
|
74 |
FIXED_EVAL_SUFFIX = """
|
|
|
164 |
|
165 |
|
166 |
# Default values for compatible mode
|
167 |
+
DEFAULT_EVAL_CRITERIA = """Does the model provide relevant and useful responses to the user's needs or questions?"""
|
168 |
|
169 |
+
DEFAULT_SCORE_1 = "The model's responses are irrelevant or unhelpful to the user's needs or queries."
|
170 |
|
171 |
+
DEFAULT_SCORE_2 = "The model sometimes provides helpful information, but often fails to address the user's actual needs or questions."
|
172 |
|
173 |
+
DEFAULT_SCORE_3 = "The model generally provides helpful responses that address the user's needs, though it may occasionally miss the mark."
|
174 |
|
175 |
+
DEFAULT_SCORE_4 = "The model regularly provides helpful responses that are well-aligned with the user's inquiries, with only rare inaccuracies."
|
176 |
|
177 |
+
DEFAULT_SCORE_5 = "The model consistently offers highly relevant and useful responses that perfectly cater to the user's needs and inquiries."
|
178 |
|
179 |
#**What are the Evaluator Prompt Templates based on?**
|
180 |
|