This repository contains a comprehensive notebook and tutorial on how to fine-tune the gemma-7b-it
model using qLora and Supervised Fine-tuning.
This project demonstrates the steps required to fine-tune the Gemma model for tasks like code generation. We use qLora quantization to reduce memory usage and the SFTTrainer
from the trl
library for supervised fine-tuning.
The notebook is available on my GitHub: gemma-Instruct-2b-Finetuning-on-alpaca.ipynb
Before running the notebook, ensure you have the following: