Submitted by PK_thundr t3_11rnppe in MachineLearning
We have a 20-40k budget at our lab and we are interested in training LLMs on data that is protected by HIPAA which puts restrictions on using just any cloud provider. We'd need a compute environment with 256gb vram.
Would it be better to use AWS EC2 P3 instances or Google Cloud instead of trying to build our own server for this? We could spend the budget on a local server, but would this be obsolete within 2 years once the next gen GPUs are released?
Disastrous_Elk_6375 t1_jc9g5v0 wrote
This is the best article for you right now - https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/
Follow the GPU recommendation chart, and check out the formulas for figuring out if buying or renting is worth it for you. Tim probably has you covered for what you need.