---
id: "concept-training-inference-chip-divergence"
type: "concept"
source_timestamps: ["00:03:10", "00:03:45"]
tags: ["hardware", "semiconductors"]
related: ["concept-inference-wall"]
definition: "The architectural necessity of using fundamentally different silicon for training AI models versus serving them in production."
sources: ["s17-3-model-drops"]
sourceVaultSlug: "s17-3-model-drops"
originDay: 17
---
# Training vs. Inference Chip Divergence

## Definition

The architectural necessity of using fundamentally different silicon for **training** AI models versus **serving (inference)** them in production.

## The Core Argument

A primary driver of the [[concept-inference-wall]] is the industry's continued reliance on the same hardware for both training and serving. The chips engineered to train massive frontier models are not optimized for inference, which has fundamentally different memory and compute characteristics:

- **Training** rewards raw matrix-multiply throughput across enormous batched workloads.
- **Inference** rewards low-latency, memory-compressed, per-query efficiency at unpredictable load.

Because chip roadmaps have been bent toward training requirements (the metric tech press cares about), serving complex models to end users remains prohibitively expensive.

## The Way Out

Resolving this requires new approaches to hardware and serving architectures. The speaker cites Google's **Turbo Quant** paper as an example — focused on compressing memory and serving more efficiently to make complex AI products economically viable for consumer scale. See [[quote-inference-chips]] for the speaker's blunt framing.

## Related
- [[concept-inference-wall]] — the macroeconomic consequence
- [[prereq-training-vs-inference]] — required background
- [[quote-inference-chips]] — "the chips we use to train should not be the chips we use to infer"
