Submitted by GPUaccelerated t3_yf3vtt in MachineLearning
I serve the AI industry, primarily building, configuring and selling GPU-accelerated workstations/servers and cloud instances.
Most people and companies buy and rent these things based on necessity. *You can't really dig holes effectively if you don't have a shovel kind of thing.*
I'm obviously not the only provider in the market. And I'm not one of the largest. Some choose me because I save them a lot of money and some choose me because I'm really really good at what I do(configuring and optimizing). (Yes, I'm confident enough to put that out there.)
When I'm taking care of an upgrade situation, it's usually because of one of two things.
- The hardware is outdated and needs a refresh to be able to support modern processing tools.
- The client's project is scaling and they need more compute power or VRAM (generally).
My question is there anyone (or companies) out there who actually cares to upgrade based on speed?
Like is anyone going through the upgrading process simply because they want to train their models faster(save time)? Or bring more value to their clients by having their models inference faster?
I'd like anyone's opinion on this but if you fit the description of this type of client, I'd like to know the thought process of upgrading. Whether you've been through it in the past or something you're going through now.
LastVariation t1_iu1mbuj wrote
Inference speed is important because that's what goes to production typically. If you're already waiting hours-days on training, then it probably takes an order of magnitude improvement to make the investment worth it. As a private consumer, my bottleneck to upgrade is my steam library.