Machine Learning

Context Distillation

A technique where the behavior of a model prompted with detailed instructions is distilled into a model that exhibits the same behavior without the instructions.

Why It Matters

Context distillation makes models cheaper to run in production by 'baking in' the system prompt behavior so you do not need to send it with every request.

Example

Training a student model on outputs from a teacher model that was given detailed system instructions, so the student behaves correctly without needing those instructions.

Think of it like...

Like training a new employee so thoroughly that they eventually do not need to reference the manual — the guidelines become second nature.

Related Terms