Overview
Therun_dataset_evaluation
method allows you to run evaluations on a dataset using one or more evaluators. This is essential for measuring model performance and data quality.
Method Signature
Synchronous
Asynchronous
Parameters
Parameter | Type | Required | Description |
---|---|---|---|
dataset_id | str | Yes | The unique identifier of the dataset to evaluate |
evaluator_ids | List[str] | Yes | List of evaluator IDs to use for evaluation |
evaluation_name | str | No | Optional name for the evaluation run |
Returns
Returns a dictionary containing the evaluation job information and status.Examples
Basic Usage
With Custom Name
Asynchronous Usage
Multiple Evaluators
Error Handling
Common Use Cases
- Quality assurance for training datasets
- Model performance benchmarking
- Automated dataset validation
- A/B testing of different model versions
- Compliance and safety checks