Skip to content
Embedding LabsEmbedding Labs
Embedding Labs
Retour à la Recherche

LoRA: Low-Rank Adaptation of Large Language Models

Hu et al.2021

Fine-tuningEfficiencyPEFT

Résumé

LoRA proposes freezing pre-trained model weights and injecting trainable rank decomposition matrices into each layer. This approach reduces trainable parameters significantly, making fine-tuning of large models feasible on limited hardware.

Pourquoi C'est Important

  • Made fine-tuning practical on constrained hardware
  • Standard approach for adapters and model personalization
  • Major reduction in storage and compute requirements

Poser une question sur cet article

Loading chat...