Tier 4

aiag - AI Agent Analysis

AI Agent Analysis

Input: $ARGUMENTS


Step 1: Identify the Work

Break down what needs to be done into discrete tasks.

DOMAIN: [Area of work]
GOAL: [What the AI agent would accomplish]

TASKS:
1. [Task] — frequency: [how often], complexity: [low/medium/high]
2. [Task] — frequency: [how often], complexity: [low/medium/high]
3. [Task] — frequency: [how often], complexity: [low/medium/high]

Step 2: Assess Automatability

For each task, evaluate AI suitability.

HIGHLY AUTOMATABLE:
- [Task] — why: [structured input, clear rules, pattern-matchable]

PARTIALLY AUTOMATABLE:
- [Task] — AI does: [part], human does: [part]

NOT AUTOMATABLE (yet):
- [Task] — why: [requires judgment, novel situations, physical presence, trust]

Criteria for automatability: structured inputs, clear success criteria, tolerance for occasional errors, availability of training data or examples.


Step 3: Evaluate Reliability Requirements

ERROR TOLERANCE BY TASK:
- [Task]: [high tolerance — errors are cheap to fix]
- [Task]: [low tolerance — errors cause real damage]

FAILURE CONSEQUENCES:
- If AI gets [task] wrong: [what happens]
- Worst case scenario: [describe]
- Recovery cost: [low/medium/high]

Step 4: Design Human Oversight

OVERSIGHT MODEL:
- [Task]: [No oversight / spot-check / review before action / human-in-the-loop]

ESCALATION TRIGGERS:
- AI should escalate when: [condition 1]
- AI should escalate when: [condition 2]
- AI should stop when: [condition]

FEEDBACK LOOP:
- How humans correct AI: [mechanism]
- How AI improves from corrections: [mechanism]

Step 5: Plan for Failure Modes

FAILURE MODES:
1. [Mode: e.g., hallucination, drift, edge case] — likelihood: [low/medium/high]
   Mitigation: [approach]
2. [Mode] — likelihood: [low/medium/high]
   Mitigation: [approach]
3. [Mode] — likelihood: [low/medium/high]
   Mitigation: [approach]

GRACEFUL DEGRADATION:
- If AI is unavailable: [fallback plan]
- If AI quality drops: [detection method and response]

Step 6: Capability Projection

CURRENT STATE: [What AI can do today for this use case]
6-MONTH PROJECTION: [What's likely to improve]
LONG-TERM OUTLOOK: [What changes if AI capabilities grow significantly]

RECOMMENDED APPROACH:
1. Start with: [lowest-risk, highest-value automation]
2. Expand to: [next tier once trust is established]
3. Monitor: [key metrics to track AI performance]

Integration

Use with:

  • /indv -> Analyze the human roles alongside AI
  • /tmsk -> Design human-AI team structure
  • /vldt -> Validate the AI agent’s outputs
  • /bldk -> Build the AI agent system incrementally