from rubric import Rubric client = Rubric() evaluation = client.evaluations.create( name="Triage Accuracy - Weekly", project="proj_abc123", dataset="ds_xyz789", evaluators=[ { "type": "triage_accuracy", "config": { "severity_weights": { "under_triage": 5.0, "over_triage": 1.0 } } }, { "type": "red_flag_detection", "config": { "protocols": ["chest_pain", "headache"] } } ], metadata={ "triggered_by": "ci_pipeline", "model_version": "v2.4.1" } ) print(f"Created evaluation: {evaluation.id}") print(f"Status: {evaluation.status}")
{ "id": "eval_def456", "object": "evaluation", "name": "Triage Accuracy - Weekly", "project": "proj_abc123", "dataset": "ds_xyz789", "status": "pending", "evaluators": [ { "type": "triage_accuracy", "config": { "severity_weights": { "under_triage": 5.0, "over_triage": 1.0 } } }, { "type": "red_flag_detection", "config": { "protocols": ["chest_pain", "headache"] } } ], "progress": { "total": 0, "completed": 0, "failed": 0 }, "created_at": "2024-01-15T10:30:00Z", "started_at": null, "completed_at": null, "metadata": { "triggered_by": "ci_pipeline", "model_version": "v2.4.1" } }
Create a new evaluation to assess a dataset using one or more evaluators.
write
Bearer gr_live_xxxxxxxx
Show evaluators properties
triage_accuracy
red_flag_detection
custom
eval_def456
evaluation
pending
running
completed
failed
cancelled
Show progress properties