Viewing a single comment thread. View all comments

trajo123 t1_iyciqd9 wrote

For users, it's quite expensive that Nvidia has such a monopoly on ML/DL compute acceleration. People replying with "don't bother, just use Nvidia&CUDA" only make the problem worse ...music for Nvidia's ears.
I would say, by all means try it out and share your experience, just be aware that it's likely going to be more hassle than using Nvidia&CUDA.

33

r_linux_mod_isahoe t1_iyclnxu wrote

no, it's AMD who fucked up. The whole ROCm is an afterthought. Hire a dev, make pytorch work on all modern AMD GPUs, then we'll talk. For now this is somehow a community effort.

26

serge_cell t1_iycp9ri wrote

For that first AMD had to make normal implementation of OpenCL. People complain all the time - slowdowns, crashes, lack of portability. This going on for 10 years already and it doesn't get better.

9

Ronny_Jotten t1_iyd43te wrote

> People replying with "don't bother, just use Nvidia&CUDA" only make the problem worse

No, they don't "only make it worse". It's good advice to a large proportion of people who just need to get work done. AMD/Intel need to hear that, and step up, by providing real, fully-supported alternatives, not leaving their customers to fool around with half-working CUDA imitations. ML is such an important field right now, and they've dropped the ball.

18

ReginaldIII t1_iycqo8c wrote

> People replying with "don't bother, just use Nvidia&CUDA" only make the problem worse ...music for Nvidia's ears.

My job is to get a trained model out the door so we can run experiments.

My job is not revolutionize the frameworks and tooling available so that competing hardware can be made a feasible alternative for everyone.

There are only so many hours in the day. I get paid for a very specific job. I have to work within the world that exists around me right now.

15

philthechill t1_iycso44 wrote

If you’re in a commercial setting your job is to get market-beating learning done at minimal cost. OP says these things might have a revolutionary cost per learning value, so yeah it is within your job parameters to look at specs, pricing and tool support at the very least. Ignoring technological revolutions is definitely one way companies end.

−7

ReginaldIII t1_iyctmv8 wrote

90% of the time my job is to be a small and consistently performing cog in a much bigger machine because I am there to help drive down stream science outcomes for other scientists (often in a different discipline).

We need to get X done within Y timeframe.

> "Lets consider upending our infrastructure and putting millions of pounds worth or existing and battle proven code and hardware up in flux so we can fuck around seeing if Intel has actually made a viable GPU-like product on their umpteenth attempt"

... is not exactly an easy sell to my board of governance.

I was in the first wave of people who got access to Xeon Phi Knights Corner Co-Processor cards. Fuck my life did we waste time on that bullshit. The driver support was abysmal, even with Intels own ICC compiler and their own MPI distribution.

11

philthechill t1_iyctys1 wrote

Yeah fair.

2

ReginaldIII t1_iydnekt wrote

Also worth considering how many years it is going to take to offset the sizeable cost of such a migration.

Forget the price of the hardware, how long is it going to take to offset the cost of the programming and administration labour to pull off this sort of move?

What about maintenance? We've got years of experience with Nvidia cards in datacentres, we understand the failure modes pretty well, we understand the tooling needed to monitor and triage these systems at scale.

What guarantees do I have that if I fill my racks with this hardware they won't be dying or catching on fire within a year?

What guarantees do I have that Intel won't unilaterally decide this is a dead cat for them and they want to scrap the project? Like they have for almost every GPU adjacent project they've had.

3

AtomKanister t1_iycydix wrote

>"Lets consider upending our infrastructure and putting millions of pounds worth or existing and battle proven code and hardware up in flux so we can fuck around seeing if Intel has actually made a viable GPU-like product on their umpteenth attempt"

That's exactly how innovation is made, and missing out on this in crucial moments is how previously big players become irrelevant in the blink of an eye. See: Kodak, Blockbuster, Sears, Nokia.

It's valid to be skeptical of new developments (because a lot of them will be dead ends), but overdo it and you're setting yourself up for disaster.

−9

hgoel0974 t1_iyczwvn wrote

Setting up infrastructure that relies on a GPU that can't do what you need yet and is not optimized for it either is certainly innovative but not in the way that you're thinking.

6

ReginaldIII t1_iydl00l wrote

> That's exactly how innovation is made

It's also how companies overextend and go out of business.

1

hgoel0974 t1_iycyx88 wrote

For most users of ML frameworks, results take priority and there isn't much they can do about AMD's shit software and unreliable support. Plus even 4090s aren't really that expensive relative to what ML people make.

That said, Intel might actually be able to compete once their drivers have caught up. Unlike AMD, who seems to have systemic issues (not to mention fatal design flaws in ROCm in general), Intel just needs time because they clearly rushed the devices out before the drivers were fully ready.

1

slashdave t1_iydpou1 wrote

Power costs dwarf hardware costs, by miles. Come up with an power efficient GPU, and we'll talk.

1