Submitted by Scarlet_pot2 t3_1277dw4 in singularity
TheKnifeOfLight t1_jed4kdv wrote
I do believe there is an ai that can run on consumer tech and can be trained on consumer tech that doesn’t need high end specs for the smaller models. I think it’s Alpaca/LLaMA and is comparable to Chatgpt from what I’m seeing online
smokingthatosamapack t1_jed6l4a wrote
the primary concern is training it to improve it. Sure fine tuning can be done but to make a substantial AI with significant changes and a unique model is its own feat that needs lots of funding.
Scarlet_pot2 OP t1_jed7tts wrote
Fine-tuning isn't the problem.. if you look at the alpaca paper, they fine tuned the LLaMA 7B model on gpt-3 and achieved gpt-3 results with only a few hundred dollars. The real costs are the base training of the model, which can be very expensive. Also having the amount of compute to run it after is an issue too.
Both problems could be helped if there was a free online system to donate compute and anyone was allowed to use it
smokingthatosamapack t1_jedmdli wrote
Yeah I see what you mean and it could happen but there's no such thing as a free lunch and even if there was a system it would probably pale in comparison to paid solutions for compute
Scarlet_pot2 OP t1_jed67k5 wrote
True alpaca is competent, but we need more models, better and larger models.. a distributed system where people donate compute could also be used to allow people to run larger models. maybe not 175 billion parameters, but maybe 50-100B as long as everyone donating compute isn't using it at the same time
that being said more smaller models like alpaca / LLaMA are needed too. if we made sufficient resources / training available to anyone, models like that could be created and made available more often
Akimbo333 t1_jefh3f9 wrote
Llama just needs much more training and fine-tuning and it'll be good
Viewing a single comment thread. View all comments