Road to Efficient LLMs 2: QLoRA
QLoRA: Efficient Finetuning of Quantized LLMs
Introduction
Previously, we discussed Low-Rank Adapters (LoRA) as a method for efficiently fine-tuning large language models (LLMs). In this post, we will discuss QLoRA, a new quantization method t...