Professional work-related project
In this project, I have provided code and a Colaboratory notebook that facilitates the fine-tuning process of an Alpaca 3B parameter model originally developed at Stanford University. The particular model that is being fine-tuned has around 3 billion parameters, which is one of the smaller Alpaca models.
The model uses low-rank adaptation LoRA to run with fewer computational resources and training parameters. We use bitsandbytes to set up and run in an 8-bit format so it can be used on colaboratory. Furthermore, the PEFT library from HuggingFace was used for fine-tuning the model.
Hyper Parameters:
Credit for Original Model: Xinyang (Young) Geng and Hao Liu from OpenLM Research
Fine-Tuned Model: RyanAir/alpaca-3b-fine-tuned (HuggingFace)