Training

LoRA

Quick Answer

Low-Rank Adaptation: a parameter-efficient fine-tuning method adding small trainable matrices.

LoRA adds small trainable low-rank matrices to model weights rather than fine-tuning all parameters. It dramatically reduces training memory, compute, and storage. For a model with millions of parameters, LoRA adds only thousands. LoRA achieves comparable quality to full fine-tuning with a fraction of the cost. It's practical for researchers and practitioners. LoRA adapters can be merged with models or kept separate. Multiple LoRA adapters can be combined. LoRA democratized fine-tuning by making it accessible.

Last verified: 2026-04-08

Compare models

See how different LLMs compare on benchmarks, pricing, and speed.

Browse all models →