AI Decision Reproducibility and Testing
QA and compliance teams needing to replay and validate AI decisions for consistency, fairness testing, and regulatory review.
Das Problem
Corules decisions are reproducible by design. Each evaluation stores a policy set version and a normalized input hash. Given these two values, the exact evaluation can be reproduced at any future point — even after policy updates — by specifying the historical policy version. This enables: regression testing when policy changes (run historical cases against new policy), fairness testing (replay a sample of decisions across demographic segments), and regulatory examination (prove that the policy in effect at time T produced outcome O for input I).
So löst Corules es
Corules's policy runtime evaluates structured context against compiled CEL expressions — returning ALLOW, BLOCK, or ESCALATE with a reason and audit ID.
Richtlinienbeispiel
// Replay any historical decision:
POST /v1/validate/replay
{
"audit_id": "aud_01J...",
"policy_set_version": "pset_v2.1.0" // historical version
}
// Returns identical result as original evaluationFrequently Asked Questions
Can I test a new policy version against historical inputs before publishing?
Yes. The simulation API runs proposed policy changes against historical audit data without affecting production. Preview the impact before publishing.
How does this support A/B testing of policies?
Run two policy versions against the same historical dataset. Compare outcome distributions. Publish the version with the desired risk/automation tradeoff.