Hands-On QLoRA
Learn how to fine-tune LLM on a custom dataset using QLoRA.
Let’s fine-tune Meta’s Llama 3.1 model on openai/gsm8k dataset using QLoRA.
Install the dependencies
First, let’s install the libraries required for fine-tuning. We'll be install the latest versions (at the time of writing) of the libraries.
pip3 install transformers==4.44.1pip3 install acceleratepip3 install bitsandbytes==0.43.3pip3 install datasets==2.21.0pip3 install trl==0.9.6pip3 install peft==0.12.0!pip install -U "huggingface_hub[cli]"
Line 1: We install the
transformerslibrary, which is a Hugging Face library that provides APIs and tools to download and train state-of-the-art pretrained models.Line 2: We install the
acceleratelibrary, which is designed to facilitate training deep learning models across different hardware. It enables the training and inference to be simple, efficient, and adaptable.Line 3: We install the
bitsandbyteslibrary, which is a transformers library that helps with the quantization of the model.Line 4: We install the
datasetlibrary for sharing and accessing datasets for downstream tasks.Line 5: We install the
trllibrary for training transformer models with reinforcement learning and supervised fine-tuning.Line 6: We install the
peftlibrary for parametric efficient fine-tuning of large language models for downstream tasks.Line 7: We install the Hugging Face CLI to log in and access the model and dataset from Hugging Face.
Hugging Face CLI
After installing the required libraries, it’s time to log in to the Hugging Face CLI. Hugging Face requires this step to access any model and dataset from the ...