blarg7459 t1_jbetts9 wrote on March 8, 2023 at 3:23 PM Reply to comment by CKtalon in [D] Can someone explain the discrepancy between the findings of LLaMA and Chinchilla? by __Maximum__ Doesn't that mean that if you include inference costs, and the model will be used extensively, you may actually get much better bang for your bucks by training much more than chinchilla-optimal? Permalink Parent 1
blarg7459 t1_iw94geo wrote on November 13, 2022 at 10:03 PM Reply to comment by allwordsaremadeup in [Research] Can we possibly get access to large language models (PaLM 540B, etc) like GPT-3 but no cost? by NLP2829 Here's how to run it locally https://towardsdatascience.com/run-bloom-the-largest-open-access-ai-model-on-your-desktop-computer-f48e1e2a9a32 Permalink Parent 9
blarg7459 t1_jbetts9 wrote
Reply to comment by CKtalon in [D] Can someone explain the discrepancy between the findings of LLaMA and Chinchilla? by __Maximum__
Doesn't that mean that if you include inference costs, and the model will be used extensively, you may actually get much better bang for your bucks by training much more than chinchilla-optimal?