timy2shoes t1_iunym8r wrote
We've been testing out their embeddings for transfer learning tasks and they've been performing quite well. Better than previous embeddings that we have tested. The 15B parameter model though is a pain in the ass. Getting the embeddings requires a workaround that is difficult to implement. Probably not worth it in my opinion.
nivrams_brain t1_iuo3l24 wrote
What kind of downstream tasks are you looking at?
timy2shoes t1_iuo4xa4 wrote
ML-guided protein engineering.
nivrams_brain t1_iuo8d1h wrote
Sounds cool, are you in academia or industry?
timy2shoes t1_iuoafb9 wrote
Industry
MangoGuyyy t1_iuq81od wrote
What company, I’m curious
[deleted] t1_iupief7 wrote
[deleted]
ROFLLOLSTER t1_iupsskm wrote
> requires a workaround that is difficult to implement
What workaround? I've also been working with ESM and tried the 15B parameter variant. It seemed worse than the 3B in my tests, but maybe I just missed the problem?
timy2shoes t1_iuptv7y wrote
We had to do a workaround to fit the 15b parameter model on a p3.8xlarge instance.
> I've also been working with ESM and tried the 15B parameter variant.
Huh. We’ve noticed the same thing. Interesting that others are having the same problem.
Mister_Abc t1_iur4gme wrote
First author here. We've had some indication that the 15B model may be overfit. It seemed to sightly improve on a few important metrics (casp14) which is why we included it.
Viewing a single comment thread. View all comments