---
id: "concept-negative-lift"
type: "concept"
source_timestamps: ["00:05:07", "00:05:15"]
tags: ["evaluation", "productivity", "failure-modes"]
related: ["framework-agent-evaluation", "claim-custom-gpts-fail-shared-work", "action-measure-review-burden"]
definition: "A phenomenon where the human time required to review and correct an AI's output exceeds the time saved by the automation, resulting in a net productivity loss."
sources: ["s06-openai-free-employee"]
sourceVaultSlug: "s06-openai-free-employee"
originDay: 6
---
# Negative Lift

## Definition

A phenomenon where the human time required to review and correct an AI's output exceeds the time saved by the automation, resulting in a net productivity loss.

## Mechanics

Negative lift occurs when the introduction of an AI tool actually decreases overall team productivity because the cost of reviewing and correcting the AI's output exceeds the time saved by automating the task. The speaker highlights this as a primary reason why early attempts to use Custom GPTs for shared team workflows (like customer service ticket triage) frequently failed — see [[claim-custom-gpts-fail-shared-work]].

In these scenarios, the AI would generate a draft or a triage decision, but because the system lacked deep context or made subtle errors, human workers had to spend significant time second-guessing and verifying the output. **If a human representative has to read the entire original ticket anyway to trust the AI's summary, the marginal utility of the AI is destroyed.**

## How to Avoid It

Deploy agents only on tasks with a 'known path' (see [[quote-known-path]] and [[framework-ideal-agent-target]]) and a very clear, objective standard for what constitutes a 'good' versus 'bad' output. The ultimate test is the [[framework-agent-evaluation|Time vs. Review evaluation]]: does the time saved by the automated draft definitively beat the [[action-measure-review-burden|review burden]] placed on the human operator?

If not, the team will naturally abandon the tool within weeks.

## Enrichment Notes

Independent enterprise AI research attributes ~74% of failed AI projects to a review-burden gap of this exact shape — making this concept a leading indicator of pilot collapse, not a fringe edge case.


## Related across days
- [[concept-trust-failure-hallucination]]
- [[concept-silent-failure]]
- [[concept-dark-code]]
- [[claim-klarna-intent-failure]]
