In this guide, we will show you how to run an LLM evaluator in the UI.
Create a new evaluator
Configure an LLM evaluator
Slug
for each evaluator. This slug will be used to apply the evaluator in your LLM calls, and will be used to identify the evaluator in the Logs.Slug
is a unique identifier for the evaluator. We suggest you don’t change it once you have created the evaluator.gpt-4o
and gpt-4o-mini
from OpenAI and Azure OpenAI.{{llm_output}}
: The output text from the LLM.{{ideal_output}}
: The ideal output text from the LLM. This is optional, you can add it if you want to give the LLM a reference output.Scoring rubric
for the evaluator. This is for the LLM to understand the scoring criteria.Passing score is the minimum score that the LLM output needs to achieve to be considered as a passing response.