Overview
While automated evaluators catch many issues, healthcare AI requires human oversight for edge cases, quality assurance, and regulatory compliance. Rubric provides purpose-built interfaces for clinician review.
How It Works
Automated Evaluation
AI outputs are scored by clinical evaluators (triage accuracy, red flags, etc.)
Review Triggers
Configure when cases should be flagged for human review:Reviewer Management
Adding Reviewers
Reviewer Qualifications
Configure which reviewers can assess which case types:| Reviewer Type | Qualifications | Case Types |
|---|---|---|
| Physician | MD/DO + specialty | Complex triage, imaging |
| Nurse | RN + triage certification | Standard triage calls |
| NP/PA | Advanced practice | Mid-level complexity |
| Specialist | Board certification | Specialty-specific cases |
Review Interface
Call Review
For voice triage review, clinicians see:- Audio player with waveform visualization
- Transcript with speaker labels and timestamps
- AI evaluation scores and extracted symptoms
- Grading form with keyboard shortcuts
Note Review
For clinical documentation review:- Source material (transcript, prior notes)
- Generated note with side-by-side comparison
- Accuracy highlights showing supported vs. unsupported claims
- Completeness checklist of required elements
Imaging Review
For radiology/pathology AI:- DICOM viewer with zoom, pan, window/level
- AI findings overlay with confidence scores
- Annotation tools for corrections
- Structured grading for findings accuracy
Assignment Strategies
Round Robin
Distribute cases evenly across available reviewers:Load Balanced
Account for reviewer availability and workload:Manual Assignment
Assign specific cases to specific reviewers:Review Workflow
Submitting Reviews
Review States
| State | Description |
|---|---|
pending | Awaiting assignment |
assigned | Assigned to reviewer |
in_progress | Reviewer has started |
completed | Review submitted |
escalated | Needs additional review |
Escalation
For complex cases requiring additional expertise:Metrics & Analytics
Track review program performance:Key Metrics
| Metric | Description | Target |
|---|---|---|
| Review TAT | Time from flagging to review completion | < 24 hours |
| Throughput | Reviews per reviewer per hour | 10-15 |
| Agreement | Inter-rater reliability (Cohen’s κ) | > 0.8 |
| Override Rate | % of AI decisions overturned | < 10% |
