Create an experiment
Create an experiment and start asynchronous workflow execution over a dataset.
Headers
Authorization
Bearer token. Use Bearer YOUR_API_KEY for API key auth or Bearer <JWT> for dashboard auth.
Request
Create and asynchronously run an experiment. Provide exactly one scoring path: evaluator_ids/evaluator_slugs or evaluator_workflow_ids.
dataset_id
Dataset ID to process.
workflow
Workflow tasks to run for each dataset row.
evaluator_ids
Preferred evaluator identifiers for scoring. Mutually exclusive with evaluator_workflow_ids.
evaluator_slugs
Backward-compatible alias for evaluator_ids. If both are provided, evaluator_ids takes precedence.
evaluator_workflow_ids
WorkflowVersion IDs configured for eval-only scoring. Mutually exclusive with evaluator IDs/slugs.
experiment_id
Optional client-provided experiment ID. The backend generates one when omitted.
name
Experiment name.
description
Experiment description.
span_workflow_name
Root workflow span name.
enable_tracing
Whether to create trace logs.
batch_size
Batch size for processing.
concurrency
Number of concurrent workers.
generation_method
Optional evaluation generation method override.
Response
Created experiment.
id
Experiment ID.
name
Experiment name.
status
Experiment execution status.
created_at
description
Experiment description.
dataset
Dataset ID associated with the experiment.
dataset_id
Dataset ID associated with the experiment.
dataset_name
Dataset name, when available.
workflow_count
Number of workflow steps.
progress
Execution progress percentage.
started_at
completed_at
workflow
Workflow tasks configured for the experiment.
evaluator_ids
Evaluator IDs used for scoring.
evaluator_slugs
Backward-compatible evaluator identifiers stored by the backend.
evaluator_workflow_ids
Eval-only workflow versions used for scoring.
batch_size
concurrency
enable_tracing
error_message
Failure details when status is failed.
Errors
400
Bad Request Error
401
Unauthorized Error
404
Not Found Error