What is the difference between Low-Rank Adaptation (LoRA) and full fine-tuning of large language models?

Machine Learning Hard

Machine Learning — Hard

What is the difference between Low-Rank Adaptation (LoRA) and full fine-tuning of large language models?

Key points

  • LoRA injects low-rank decomposition matrices into attention layers
  • Full fine-tuning updates all model parameters
  • LoRA reduces memory requirements
  • LoRA achieves comparable performance
  • Full fine-tuning updates all layers including attention

Ready to go further?

Related questions