this post was submitted on 28 Nov 2023
1 points (100.0% liked)
LocalLLaMA
3 readers
1 users here now
Community to discuss about Llama, the family of large language models created by Meta AI.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
With a full finetune I don't think so -- the LIMA paper showed that 1000 high quality samples is enough with a 65B model. With QLoRA and LoRA, I don't know. The number of parameters you're affecting is set by the rank you choose. It's important to get the balance between the rank, dataset size, and learning rate right. Style and structure is easy to impart, but other things not so much. I often wonder how clean the merge process actually is. I'm still learning.