Spaces:
Runtime error
Runtime error
Zeno + OpenAI Evals
OpenAI's Evals library is a great resource providing evaluation sets for LLMS.
This repo provides a hub for exploring these results using the Zeno evaluation tool.
Add New Evals
To add new evals, add a new entry to evals/evals.yaml with the following fields:
results-file: The first.jsonlresult fromoaievalslink: A link to the evals commit for this evaluationdescription: A succint description of what the evaluation is testingsecond-results-file: An optional second.jsonlresult fromoaievals. Must be the same dataset as the first one.functions-file: An optional Python file with Zeno functions for the evaluations.
Make sure you test your evals locally before submitting a PR!
Running
poetry install
python -m zeno-evals-hub evals/evals.yaml