Spaces:
Running
Running
Improved wording
Browse files
common.py
CHANGED
|
@@ -1,10 +1,8 @@
|
|
| 1 |
# Page Headers
|
| 2 |
-
MAIN_TITLE = "# Judge Arena -
|
| 3 |
-
SUBTITLE = "*Free LLM Evals to test your GenAI application.*"
|
| 4 |
|
| 5 |
# How it works section
|
| 6 |
HOW_IT_WORKS = """
|
| 7 |
-
# How it works:
|
| 8 |
- **Run any form of evaluation:** from simple hallucination detection to qualitative interpretations
|
| 9 |
- **Evaluate anything:** coding, analysis, creative writing, math, or general knowledge
|
| 10 |
"""
|
|
@@ -13,7 +11,8 @@ BATTLE_RULES = """
|
|
| 13 |
## 🤺 Battle Rules:
|
| 14 |
- Both AIs stay anonymous - if either reveals its identity, the duel is void
|
| 15 |
- Choose the LLM judge that most aligns with your judgement
|
| 16 |
-
- If both score the same - choose the critique that you prefer more
|
|
|
|
| 17 |
"""
|
| 18 |
|
| 19 |
# CSS Styles
|
|
@@ -29,6 +28,18 @@ CSS_STYLES = """
|
|
| 29 |
"""
|
| 30 |
|
| 31 |
# Default Eval Prompt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
DEFAULT_EVAL_PROMPT = """You are assessing a chat bot response to a user's input based on the helpfulness of the response.
|
| 33 |
|
| 34 |
Score:
|
|
|
|
| 1 |
# Page Headers
|
| 2 |
+
MAIN_TITLE = "# Judge Arena - Free LLM Evals to test your GenAI application"
|
|
|
|
| 3 |
|
| 4 |
# How it works section
|
| 5 |
HOW_IT_WORKS = """
|
|
|
|
| 6 |
- **Run any form of evaluation:** from simple hallucination detection to qualitative interpretations
|
| 7 |
- **Evaluate anything:** coding, analysis, creative writing, math, or general knowledge
|
| 8 |
"""
|
|
|
|
| 11 |
## 🤺 Battle Rules:
|
| 12 |
- Both AIs stay anonymous - if either reveals its identity, the duel is void
|
| 13 |
- Choose the LLM judge that most aligns with your judgement
|
| 14 |
+
- If both score the same - choose the critique that you prefer more!
|
| 15 |
+
<br><br>
|
| 16 |
"""
|
| 17 |
|
| 18 |
# CSS Styles
|
|
|
|
| 28 |
"""
|
| 29 |
|
| 30 |
# Default Eval Prompt
|
| 31 |
+
EVAL_DESCRIPTION = """
|
| 32 |
+
## 📝 Instructions
|
| 33 |
+
**Precise evaluation criteria leads to more consistent and reliable judgments.** A good evaluation prompt should include the following:
|
| 34 |
+
- Evaluation criteria
|
| 35 |
+
- Scoring rubric
|
| 36 |
+
- (Optional) Examples\n
|
| 37 |
+
|
| 38 |
+
**Any variables you define in your prompt using {{double curly braces}} will automatically map to the corresponding input fields under "Sample to evaluate" section on the right.**
|
| 39 |
+
|
| 40 |
+
<br><br>
|
| 41 |
+
"""
|
| 42 |
+
|
| 43 |
DEFAULT_EVAL_PROMPT = """You are assessing a chat bot response to a user's input based on the helpfulness of the response.
|
| 44 |
|
| 45 |
Score:
|