Viewing a single comment thread. View all comments

ggf31416 t1_j0ypnpp wrote

Training a large model only in CPU is madness, it will take forever and waste a lot of electricity. You need a GPU with CUDA or an equivalent solution fully supported by your framework. See e.g. this benchmark.

A t2.micro instance may be free during the free trial but is useless for anything resource intensive. You are much better off just using Google Colab or Kaggle notebooks.

If you have to train models very often (like everyday) and 24GB from a RTX3090 or better a RTX4090 is enough, a dedicated computer is the most cost effective way in the long run. If you cant afford a RTX3090 and 12GB is enough, a 3060 with 12GB will do (for ML we usually want as much VRAM as possible, raw computing power often is not the bottleneck).

Vast ai is a cost effective way of renting computing power for non-constant use, much cheaper than AWS or GCP, but beware that because of how it works the instance is not fully secure against attacks from the host so you can't use it with sensitive data.

Any good CUDA GPU will be able to train with a small dataset in less of a day, so take that into account for the decision between purchasing a GPU and cloud computing.

7

RealDaddy2020 t1_j0zltw6 wrote

Good read, thank you. What are your thoughts about the NVIDIA TESLA M40 24GB, would this be a good alternative over a RTX 3060?

1

ggf31416 t1_j0zw53a wrote

Not very good https://www.reddit.com/r/MLQuestions/comments/rttzxg/tesla_m40_24gb_gpu_very_poor_machinelearning

Tesla cards typically are more expensive than the consumer equivalent even at the same performance and amount of memory.

2

RealDaddy2020 t1_j1078ok wrote

Ok. What a shame. eBay are selling them for 100-130USD. Thought I found a good deal. Will look at the 3060 12gb instead.

2