Tier 4

sadrt

Divergence Risk Test

Input: $ARGUMENTS


Overview

Estimate where two independent executors would diverge when applying the same gate/procedure to the same input. If a procedure produces different outputs depending on who runs it, it’s not reliable enough to be a standard.

Steps

Step 1: Identify Divergence Triggers

For the target procedure/gate, flag potential divergence sources:

Trigger TypeWhat to Look ForExample
Undefined termsKey nouns without definitions”adequate” — adequate by whose standard?
Bundled questionsMultiple checks in one”Is it complete and correct?” — could disagree on either
Non-operational criteriaCan’t be checked mechanically”Does it feel right?”
Missing answer-interfaceNo way to look up the answer”Is the market ready?” — how to check?
Ambiguous stopping rulesNo clear “done” signal”Analyze until satisfied”
Implicit orderingSteps that could be done in different orders with different results”Consider A and B” — which first?
Subjective thresholdsNumbers without benchmarks”Enough evidence” — how much?

Step 2: Simulate Two Executors

For each trigger found, imagine two competent but independent executors:

Executor A: Interprets conservatively, applies strict standards, focuses on completeness Executor B: Interprets pragmatically, applies reasonable standards, focuses on efficiency

For each trigger:

TriggerA’s InterpretationB’s InterpretationThey Agree?
[trigger][A reads it as…][B reads it as…]Y/N

Step 3: Rate Divergence Risk

For each disagreement:

DisagreementImpactLikelihoodRisk Level
[where they diverge]How different are the outputs? (H/M/L)How likely is this divergence? (H/M/L)[H×H=Critical, etc.]

Step 4: Propose Fixes

For each divergence risk:

RiskFix TypeProposed Fix
Undefined termDefine”[term] means [specific definition]“
Bundled questionSplitSeparate into independent checks
Non-operationalOperationalizeReplace with specific, checkable criterion
Missing interfaceAdd reference”Check using [method/source]“
Ambiguous stoppingAdd rule”Stop when [specific condition]“
Subjective thresholdQuantify”At least [N] items/sources/examples”

Step 5: Validate Fixes

After proposing fixes, re-run the executor simulation:

  1. Would A and B now agree?
  2. Did the fix introduce new ambiguity?
  3. Is the fix practical (not so rigid it prevents judgment)?

Step 6: Report

DIVERGENCE RISK TEST:
Target: [what was tested]
Triggers found: [N]

Risk assessment:
| # | Trigger | Risk Level | Fix |
|---|---------|-----------|-----|
| 1 | [trigger] | [Critical/High/Medium/Low] | [proposed fix] |

Divergence hotspots: [sections with most triggers]
Estimated executor agreement: [%] before fixes, [%] after fixes

Priority fixes: [top 3 most impactful]

When to Use

  • When you want repeatability across executors (human, AI, future you)
  • When a gate/procedure is central to enforcement
  • When disagreements keep appearing in review
  • → INVOKE: /sapea (executability audit) for deeper executability check
  • → INVOKE: /saqrc (question rewrite chains) for fixing unclear questions

Verification

  • Multiple trigger types checked
  • Two-executor simulation performed
  • Divergence risk rated by impact and likelihood
  • Fixes proposed for high-risk triggers
  • Fixes validated (don’t introduce new divergence)