---
id: "concept-ai-reviewing-ai"
type: "concept"
source_timestamps: ["00:06:11", "00:07:05"]
tags: ["quality-assurance", "evaluations", "compounding-advantages"]
related: ["framework-agentic-eval-loop", "action-implement-ai-review-pipelines"]
definition: "The deployment of secondary AI systems to autonomously audit, critique, and refine the output of primary AI generators before human review."
sources: ["s35-compounding-gap"]
sourceVaultSlug: "s35-compounding-gap"
originDay: 35
---
# AI Reviewing AI (Agentic Eval Loops)

## AI Reviewing AI (Agentic Eval Loops)

One of the most **underrated compounding advantages** in the near future is using AI to review work generated by other AI.

### Paradigm shift
- **Old paradigm**: AI creates the draft, human reviews it.
- **New paradigm**: AI creates the draft, AI audits the draft, **human applies finishing touches**.

### What AI reviewers catch
Dedicated AI reviewers are deployed to catch:

- Inconsistencies
- Missed requirements
- Risky assumptions
- Bad architectural choices

### Already happening in engineering
Smart engineering teams are already building **eval loops agentically**, where code is repeatedly checked by an AI until it passes **5–8 different evaluation sets** before a human ever sees it. The procedural specification is captured in [[framework-agentic-eval-loop]].

### Generalization
This pattern will extend across **all knowledge work**, turning triage and review into highly simplified, high-leverage activities for humans.

### How to act on this
See [[action-implement-ai-review-pipelines]].

### Enrichment context
Evaluation-as-a-Service vendors (Scale AI, Honeycomb) already operationalize multi-metric AI-to-AI review in production. This prediction is the **least speculative** of the ten.


## Related across days
- [[concept-meta-task-agent-split]]
- [[framework-agentic-eval-loop]]
- [[action-build-deterministic-evals]]
- [[concept-comprehension-gate]]
