Submitted by austintackaberry t3_120usfk in MachineLearning
machineko t1_jdnmg8l wrote
Reply to comment by ephemeralentity in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Right, 8GB won't be enough for LLaMA 7b. You should try GPT-2 model. That should work on 8GB VRAM.
ephemeralentity t1_jdp2pu8 wrote
Thanks looks like gpt2 worked! Sorry, stupid question but how do I save/re-use the results of my model finetune? When I re-finetune for 0:2 epochs it gives a reasonable response but if I try to skip model.finetune, it responds with new lines only (\n\n\n\n\n\n\n\n ...).
machineko t1_jdqzmyq wrote
model.save("path/to/your/weights") saves it to the directory
After that, you can load it with
model = BaseModel.create("gpt2", "path/to/your/weights")
Can you share the input text you have used? It is possible that GPT-2 is too small and needs custom generation parameters.
ephemeralentity t1_jdt1krp wrote
Thanks a lot! To be honest, I need to spend a bit more time familiarising myself with pytorch / this package. I'll see if I can figure it out from here.
machineko t1_jdtv8jv wrote
If you need help, come find us on our discord channel.
Viewing a single comment thread. View all comments