Run LLM Evals

Run online evals for your test cases using the metrics in metricCollection.

Headers

CONFIDENT_API_KEYstringRequired
The API key of your Confident AI project.

Request

metricCollectionstringRequired
The name of the metric collection you wish to use for evaluation.
llmTestCaseslist of objectsOptional

This is a list of single-turn test cases to evaluate. If you are evaluating multi-turn test cases, this should be null.

conversationalTestCaseslist of objectsOptional

This is a list of multi-turn test cases to evaluate. If you are evaluating single-turn test cases, this should be null.

hyperparametersmap from strings to anyOptional
This is any hyperparameters like model or prompt you wish to associate with the test run.
identifierstringOptional
A unique identifier for the test run.

Response

This endpoint returns an object.
successboolean
This is true if the test cases were successfully evaluated.
dataobject
deprecatedboolean
This is true if this endpoint is deprecated.