Submitted by Zealousideal-Copy463 t3_10khmxo in deeplearning
So I've been wondering for a while if maybe I should get a 4090, or if I should just use AWS or something.
For context: I work at a tech company and we use tensorflow/pytorch so I have a decent experience with that. I have used mostly AWS to train and test things. The problem is, in my experience, moving data from S3 to Sagemaker is a pain in the ass, and I have only used like 1-2 GB of data, mostly tabular data.
Now I want to test a few things myself, train some image models. I've been playing with some models and I got 100 GB of data that I want to fit a model with. I have tried with Colab and the data in google drive, but drive gets confused with multiple files so it's really annoying.
Any suggestions on how to do this in the cloud? I also have some experience with GCP and Azure, but AWS is the provider I have the most experience with. Can I do this without suffering too much when handling data around or should I just buy a 4090 and train stuff locally?
agentfuzzy999 t1_j5qy82t wrote
“Should I just buy a 4090”
Ok Jeff Bezos
4090 clock speed is going to be faster than similar instances that use T4s, plus wayyyyyyy more CUDA cores. Training will be significantly faster, if you can fit the model on the 4090. If you can “business expense” a 4090 for your own machine, good lord do that.