Submitted by GPUaccelerated t3_yf5jm3 in deeplearning
I serve the AI industry, primarily building, configuring and selling GPU-accelerated workstations/servers and cloud instances.
Most people and companies buy and rent these things based on necessity. *You can't really dig holes effectively if you don't have a shovel kind of thing.*
I'm obviously not the only provider in the market. And I'm not one of the largest. Some choose me because I save them a lot of money and some choose me because I'm really really good at what I do(configuring and optimizing). (Yes, I'm confident enough to put that out there.)
When I'm taking care of an upgrade situation, it's usually because of one of two things.
- The hardware is outdated and needs a refresh to be able to support modern processing tools.
- The client's project is scaling and they need more compute power or VRAM (usually).
My question is there anyone (or companies) out there who actually cares to upgrade based on speed?
Like is anyone going through the upgrading process simply because they want to train their models faster(save time)? Or bring more value to their clients by having their models inference faster?
I'd like anyone's opinion on this but if you fit the description of this type of client, I'd like to know the thought process of upgrading. Whether you've been through it in the past or something you're going through now.
sckuzzle t1_iu2aa7o wrote
We use models to control things in real-time. We need to be able to predict what is going to happen in 5 or 15 minutes and proactively take actions NOW. If it takes 5 minutes to predict what is going to happen 5 minutes in the future, the model is useless.
So yes. We care about speed. The faster it runs the more we can include in the model (making it more accurate).