Submitted by austintackaberry t3_120usfk in MachineLearning
ephemeralentity t1_jdp2pu8 wrote
Reply to comment by machineko in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Thanks looks like gpt2 worked! Sorry, stupid question but how do I save/re-use the results of my model finetune? When I re-finetune for 0:2 epochs it gives a reasonable response but if I try to skip model.finetune, it responds with new lines only (\n\n\n\n\n\n\n\n ...).
machineko t1_jdqzmyq wrote
model.save("path/to/your/weights") saves it to the directory
After that, you can load it with
model = BaseModel.create("gpt2", "path/to/your/weights")
Can you share the input text you have used? It is possible that GPT-2 is too small and needs custom generation parameters.
ephemeralentity t1_jdt1krp wrote
Thanks a lot! To be honest, I need to spend a bit more time familiarising myself with pytorch / this package. I'll see if I can figure it out from here.
machineko t1_jdtv8jv wrote
If you need help, come find us on our discord channel.
Viewing a single comment thread. View all comments