Quantized Low-Rank Adaptation (QLoRA)
Learn about the components and working of the Quantized Low-Rank Adaptation (QLoRA) technique.
Quantized Low-Rank Adaptation (QLoRA), as the name suggests, combines the two most widely used methods of fine-tuning, i.e., LoRA and quantization. Where LoRA uses the low-rank matrices to reduce the number of trainable parameters, QLoRA extends it by further reducing the model size by quantizing its weights.
Get hands-on with 1400+ tech skills courses.