Submitted by jaxolingo t3_125qztx in MachineLearning
LetGoAndBeReal t1_je9zfyb wrote
Reply to comment by Goldenier in [D] The best way to train an LLM on company data by jaxolingo
>And there is no reason why the same methods wouldn't work on LLMs too, for example there is already Lora for LLMs too.
It's really not helpful to make strong assertions like this without referring to specific, verifiable sources. Fine-tuning very typically is done in a way where certain layers/parameters of the model are frozen. This is done to avoid the sort of loss we are discussing. The LoRA paper itself states that LoRA "freezes the pre-trained model weights".
Viewing a single comment thread. View all comments