Viewing a single comment thread. View all comments

EmbarrassedHelp t1_j6ixyl6 wrote

LAION and other groups are working on open source chatbots right now as we speak, and they're making great progress.


NoUtimesinfinite t1_j6j00k1 wrote

The problem isnt training. If initial upfront cost was the only barrier then yes a free version would eventually pop up. The problem is that each query costs a lot, something that cannot be made up by ad rev so anyone running the servers will require money to run it.


GreatBigJerk t1_j6j01xh wrote

It's extremely unlikely that regular people will be able to run anything close to as good as ChatGPT for several years. Language models are far more resource hungry than things like Stable Diffusion.


neoplastic_pleonasm t1_j6j72be wrote

The ChatGPT model is in the neighborhood of 750GB, so sadly we won't be seeing anything remotely as capable that can run on consumer hardware any time soon.


slashd t1_j6jhrth wrote

> 750GB

That easily fits on a $50 1TB ssd 😁


neoplastic_pleonasm t1_j6jk8gt wrote

Yep, now you only need a hundred thousand dollars more for a GPU cluster with enough VRAM to run inference with it.


NegotiationFew6680 t1_j6jmsiq wrote


Now imagine how slow that would be.

There’s a reason these models are run on distributed clusters. A single query to ChatGPT is likely being processed by multiple GPUs across dozens of machines


gmes78 t1_j6k6myq wrote

You need to fit it in GPU VRAM. So go ahead and show me a consumer GPU with 750GB of VRAM.