Skip to content

Why is it recommend to set load_in_8bit: true for LORA finetuning? #1611

Closed Answered by NanoCode012
rudolpheric asked this question in Q&A
Discussion options

You must be logged in to vote

Hey, this reply is a bit late, but I hope I can clarify this for future readers.

The reason is that when you do LORA fine-tuning, you're fine-tuning a separate set of weights. It is recommended that you load the base model in 8bit to save vram which allows you to set higher batch size + train faster.

and still the model always gets worse through lora finetuning

Could this be an issue with the dataset?

Shouldn't the model loose performance through quantisation?

Once merged, your model will be outputted as fp16.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by NanoCode012
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants