WarProfessional3278
WarProfessional3278 t1_je790g7 wrote
Reply to [D] Training a 65b LLaMA model by Business-Lead2679
By training do you mean finetuning with lora or from the ground up like alpaca? Realistically you could just rent an 8xa100 and spend 4 or 5 hours to get it done
WarProfessional3278 t1_je6y0xm wrote
Reply to comment by __god_bless_you_ in We are opening a Reading Club for ML papers. Who wants to join? 🎓 by __god_bless_you_
RLHF, sparks of AGI, or emergent in LLM.
The papers listed here are important foundational works that are outdated and improved beyond recognition by current sota.
WarProfessional3278 t1_je6mjh0 wrote
I like the idea but you guys should really update that to-read list. Most of these papers are made obsolete in the last year or this.
WarProfessional3278 t1_je1w1xo wrote
Reply to comment by BlackstockTy476 in GPT's Language Interpretation will make traveling so much better by BlackstockTy476
>It was shown to do translations better than Google and DeepL
Oh really? Do you know a good benchmark for that?
WarProfessional3278 t1_je1giqi wrote
Bad demo imo. The mom's Chinese is worse than English->Chinese google translate.
Plus, there are way too many tools out there that does real time speech to speech translation better than GPT.
WarProfessional3278 t1_jdrzo00 wrote
Reply to [D] GPT4 and coding problems by enryu42
Horace He made a nice thread on this when GPT-4 first came out. Realistically this is expected - within the short time span, there isn't much else you can do to improve the model performance other than increasing size of training data, which resulted in data contamination.
I expect the next "big thing" to be some of self-correcting output, or better chain-of-thoughts reasoning.
WarProfessional3278 t1_jaj9nnt wrote
Reply to comment by harharveryfunny in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Rough estimate: with one 400w gpu and $0.14/hr electricity, you are looking at ~0.00016/sec here. That's the price for running the GPU alone, not accounting server costs etc.
I'm not sure if there are any reliable estimate on FLOPS per token inference, though I will be happy to be proven wrong :)
WarProfessional3278 t1_j649od6 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Does anyone know of any good AI-generated text detectors? I know there's GPTZero but it's not very good in my experience.
My research has led me to Hive AI but I'm sure there are better alternatives out there that does not claim such good results (99.9% accuracy) while still having a lot of false positives in my tests.
WarProfessional3278 t1_je8i1a9 wrote
Reply to OPUS AI: Text-to-Video Game, the future of video gaming where you type and a 3D World emerges: A Demo by Hybridx21
Just a heads up, this demo has existed for more than two years. Here's the original demo by the CEO of said company (posted on Jun. 23, 2020). Also, the tweet author linked here is a pretty unreliable source imo.
I have been unable to find any playable alpha version of their software, so I have to remain skeptical of how it actually works. The demo could be scripted.