KBM_KBM
KBM_KBM t1_ja70ytj wrote
Hopefully it is easier to use than pytorch geometric
KBM_KBM t1_j5hnoke wrote
Reply to [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
Maybe say if we can download our answer then in the file we get the answer some watermark is encoded.
KBM_KBM t1_j4j10y6 wrote
You can pre train and finetune energy efficient language models such as electra or convbert in this gpu. But maybe the batch size might not be too big so the descent would be a bit noisy and also keep the corpus size as small as possible.
Look into bio electra paper which also has the notebook on how he has trained it .
KBM_KBM t1_j3gere2 wrote
Reply to comment by singularpanda in [D] Will NLP Researchers Lose Our Jobs after ChatGPT? by singularpanda
True but practically training a gpt model is not computationally cheap. I think instead of making such generalized language models we need to focus more one subject specific language models.
KBM_KBM t1_j3g7swj wrote
Reply to comment by singularpanda in [D] Will NLP Researchers Lose Our Jobs after ChatGPT? by singularpanda
https://github.com/lucidrains/PaLM-rlhf-pytorch
Similar to chat get architecture you can play with this
KBM_KBM t1_j35r12q wrote
Do you have target labels for how much defect do you have against the flow?
KBM_KBM t1_iyeemzd wrote
Reply to [D] Other than data what are the common problems holding back machine learning/artificial intelligence by BadKarma-18
I believe lack of privacy preserving models/ systems in machine learning is also a cause for concern. In many applications we collect user data (the biggest companies in the world do that) if we are able to train models in a more secure manner(not just federated learning) people might be more inclined to give more info if they can control and we can guarantee no party apart from the user will get a peek at the raw data.
This is my two cents on the topic.
KBM_KBM t1_ja71zrh wrote
Reply to [D] Is RL dead/worth researching these days? by [deleted]
Chat gpt works using a combination of rl and llm