Training

Instruction Tuning

Quick Answer

Fine-tuning on instruction-following examples to teach models to follow user directions.

Instruction tuning fine-tunes models on examples of instructions paired with high-quality responses. This transforms raw language models into helpful assistants. Instruction tuning teaches models to: follow explicit instructions, refuse unsafe requests, and format responses appropriately. High-quality instruction data is more valuable than quantity. Instruction tuning is typically followed by RLHF or DPO for further improvement. Many open models use instruction tuning. It's foundational for making models practical.

Last verified: 2026-04-08

Compare models

See how different LLMs compare on benchmarks, pricing, and speed.

Browse all models →