Spaces:
Runtime error
Runtime error
Commit
Β·
1ca62af
1
Parent(s):
406a5d4
README
Browse files
README.md
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
---
|
| 2 |
title: Zeno Evals Hub
|
| 3 |
emoji: π
|
| 4 |
colorFrom: pink
|
|
@@ -7,9 +7,25 @@ sdk: docker
|
|
| 7 |
pinned: false
|
| 8 |
license: mit
|
| 9 |
fullWidth: true
|
| 10 |
-
---
|
| 11 |
|
| 12 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
|
| 14 |
### Running
|
| 15 |
|
|
|
|
| 1 |
+
<!-- ---
|
| 2 |
title: Zeno Evals Hub
|
| 3 |
emoji: π
|
| 4 |
colorFrom: pink
|
|
|
|
| 7 |
pinned: false
|
| 8 |
license: mit
|
| 9 |
fullWidth: true
|
| 10 |
+
--- -->
|
| 11 |
|
| 12 |
+
# Zeno + OpenAI Evals
|
| 13 |
+
|
| 14 |
+
OpenAI's [Evals library](https://github.com/openai/evals) is a great resource providing evaluation sets for LLMS.
|
| 15 |
+
|
| 16 |
+
This repo provides a hub for exploring these results using the [Zeno](https://zenoml.com) evaluation tool.
|
| 17 |
+
|
| 18 |
+
## Add New Evals
|
| 19 |
+
|
| 20 |
+
To add new evals, add a new entry to `evals/evals.yaml` with the following fields:
|
| 21 |
+
|
| 22 |
+
- `results-file`: The first `.jsonl` result from `oaievals`
|
| 23 |
+
- `link`: A link to the evals commit for this evaluation
|
| 24 |
+
- `description`: A succint description of what the evaluation is testing
|
| 25 |
+
- `second-results-file`: An optional second `.jsonl` result from `oaievals`. Must be the same dataset as the first one.
|
| 26 |
+
- `functions-file`: An optional Python file with [Zeno functions](https://zenoml.com/docs/api) for the evaluations.
|
| 27 |
+
|
| 28 |
+
Make sure you test your evals locally before submitting a PR!
|
| 29 |
|
| 30 |
### Running
|
| 31 |
|