Evaluations
Run models against your data
Introducing Evaluations, a powerful feature designed to enable you to effortlessly test and compare a selection of AI models against your datasets.
Whether you're fine-tuning models or evaluating performance metrics, Oxen evaluations simplifies the process, allowing you to quickly and easily run prompts through an entire dataset.
Once you're happy with the results, output the resulting dataset to a new file, another branch, or directly as a new commit.
Summarizing Legal Documents
d9f378d3-92f1-4b9b-b21a-54a579589208
Running started
Mathias Barragan
Mathias Barragan
1 month ago
Prompt: You are a lawyer specializing in simplifying legal documents. You will read the following documents, analyze and think about them, then output a summarization containing only the necessary information we need in simplified terms. Do not output anything else besides the summarization. Here is the document: {text}
1 iteration$ 0.0000
texttextopenaiOpenAI/o1 preview
First Run with Llama 3.1 8B
7977b51a-c2a5-4a89-b57f-7a2290698922
2 / 5 row sample cancelledcancelled
Mathias Barragan
Mathias Barragan
1 month ago
Prompt: text
1 iteration 8951 tokens$ 0.0012
textembeddingsopenaiOpenAI/Text Embedding 3 - Large