Artificial Intelligence

Scaling Laws

Empirical findings showing predictable relationships between model performance and factors like model size (parameters), dataset size, and compute budget. Performance improves as a power law with these factors.

Why It Matters

Scaling laws guide the multi-billion dollar decisions about how much to invest in training AI models. They predict performance before spending the compute.

Example

Doubling the model size from 7B to 14B parameters might reduce loss by ~15%, and this relationship holds predictably across many orders of magnitude.

Think of it like...

Like the economics of factory production — there are reliable rules about how output improves as you invest more in equipment, workers, and raw materials.

Related Terms