Run LLM Evals
Run online evals for your test cases using the metrics in metricCollection.
Headers
CONFIDENT_API_KEY
The API key of your Confident AI project.
Request
metricCollection
The name of the metric collection you wish to use for evaluation.
llmTestCases
This is a list of single-turn test cases to evaluate. If you are evaluating multi-turn test cases, this should be null.
conversationalTestCases
This is a list of multi-turn test cases to evaluate. If you are evaluating single-turn test cases, this should be null.
hyperparameters
This is any hyperparameters like model or prompt you wish to associate with the test run.
identifier
A unique identifier for the test run.
Response
This endpoint returns an object.
success
This is true if the test cases were successfully evaluated.
data
deprecated
This is true if this endpoint is deprecated.