Viewing a single comment thread. View all comments

wendten t1_jdfh6ya wrote

best is a very vague term. Do you have access to a gpu cluster, or do you plan to run it on an office laptop. However id say the Alpaca model would be a good candidate. you can follow their guidance and make your own custom model from one of metas Llama models

7

sweatierorc t1_jdfwh5f wrote

LLama is not open-source (though it is gratis).

Edit: typo

6

ironmagnesiumzinc t1_jdfjdbj wrote

When fine-tuning alpaca, would it make sense to use an unsupervised approach with raw text data (say only 100mb) or would a supervised QA approach be way better?

3

to4life4 OP t1_jdfhhup wrote

"Best" I suppose meaning closest to the latest ChatGPT on the usual benchmarks.

First on my own gpu to test (3080ti), then on a cluster if I can prove out the concept.

Thanks I'll definitely look into Alpaca. It can be customized to work with human ratings of generated output?

1