315

Sorry if I'm not the first to bring this up. It seems like a simple enough solution.

you are viewing a single comment's thread
view the rest of the comments
[-] JackGreenEarth@lemm.ee 29 points 9 months ago* (last edited 9 months ago)

What other company besides AMD makes GPUs, and what other company makes GPUs that are supported by machine learning programs?

[-] meteokr@community.adiquaints.moe 25 points 9 months ago

Exactly, Nvidia doesn't have real competition. In gaming sure, but no one is actually competiting with CUDA.

[-] jon@lemmy.tf 9 points 9 months ago

AMD has ROCm which tries to get close. I've been able to get some CUDA applications running on a 6700xt, although they are noticeably slower than running on a comparable NVidia card. Maybe we'll see more projects adding native ROCm support now that AMD is trying to cater to the enterprise market.

[-] Turun@feddit.de 5 points 9 months ago

They kinda have that, yes. But it was not supported on windows until this year and is in general not officially supported on consumer graphics cards.

Still hoping it will improve, because AMD ships with more VRAM at the same price point, but ROCm feels kinda half assed when looking at the official support investment by AMD.

[-] meteokr@community.adiquaints.moe 3 points 9 months ago

I don't own any nvidia hardware out of principal, but ROCm is no where even close to cuda as far as mindshare goes. At this point I rather just have a cuda->rocm shim I can use, in the same was as directx->vulkan does with proton. Trying to fight for mindshare sucks, so trying to get every dev to support it just feel like a massive uphill battle.

[-] dudewitbow@lemmy.ml 11 points 9 months ago

AMD supports ML, its just a lot of smaller projects are made with CUDA backends, and dont have developers there to switch from CUDA to OpenCL or similar.

Some of the major ML libraries that used to built around CUDA like Tensorflow has already made non CUDA branches, but thats only because tensorflow is open source, ubiquitous in the scene and litterally has google behind it.

ML for more niche uses basically is in the chicken and egg situation. People wont use other gpus for ML because theres no dev working on non CUDA backends. No ones working on non CUDA backends because the devs end up buying Nvidia, which is basically what Nvidia wants.

There are a bunch of followers but a lack in of leaders to move the direction in a more open compute environment.

[-] PlatinumSf@pawb.social 2 points 9 months ago

Huh, my bad. I was operating off of old information. They've actually already released the sdk and apis I was referring to.

[-] coffeetest@kbin.social 10 points 9 months ago

My Intel Arc 750 works quite well at 1080 and is perfectly sufficient for me. If people need hyper refresh rates and resolution and all all the bells well then have fun paying for it. But if you need functional, competent gaming, at US$200 Arc is nice.

[-] PlatinumSf@pawb.social 6 points 9 months ago

No joke, probably intel. The cards won't hold a candle to a 4090 but they're actually pretty decent for both gaming and ML tasks. AMD definitely needs to speed up the timeline on their new ML api tho.

[-] JoeCoT@kbin.social 6 points 9 months ago

Problem with Intel cards is that they're a relatively recent release, and not very popular yet. It's going to be a while before games optimize for them.

For example, the ARC cards aren't supported for Starfield. Like they might run but not as well as they could if Starfield had optimized for them too. But the card's only been out a year.

[-] luna@lemmy.catgirl.biz 3 points 9 months ago

The more people use Arc the quicker it becomes mainstream and optimised for but arc is still considered "beta" and slow in peoples minds even though there were huge improvements and the old benchmarks don't hold any value anymore. chicken and Egg problem. :/

Disclaimer: i have an arc 770 16GB because every other sensible upgrade path would have cost 3x-4x more for the same performance uplift (and I'm not buying an 8GB card in 2023+) but now I'm starting to get really angry at people blaming Intel for "not supporting this new game" - all that gpus should support is the graphics API to the letter of the specification, all this day-1 patching and driver hotfixes to make games run decent is bs. Games need to feed the API and GPUs need to process what the API tells it to, nothing more nothing less. It's a complex issue and i think Nvidia held the monopoly for too long, everything is optimised for Nvidia at the cost of making it worse for everyone else.

[-] dan@upvote.au 6 points 9 months ago* (last edited 9 months ago)

Isn't the entire point of DirectX and OpenGL that it abstracts away the GPU-specific details? You write code once and it works on any graphics card that supports the standard? It sounds like games are moving towards what we had in the old days, where they have specific code per graphics card?

[-] luna@lemmy.catgirl.biz 4 points 9 months ago

I think the issue started with gpu-architecture tailored technologies like physx or gameworks but im probably wrong. For example I have nothing against physx but it only runs on nvidia cores natively (fast), i have an issue when there's a monetary incentive or exclusive partnering of nvidia and game studios - so if you want to play the game with all the features, bells and whistles, it was designed with you would need to also buy their overpriced (and current gen: underperforming) gpus just because you'd be missing out on features or performance on any other gpu architecture.

If this trend continues everybody will need a €1k+ gpu from nvidia and a €1k+ gpu from AMD and hot-swap between them depending on what game you wish to play.

[-] Erdrick@beehaw.org 3 points 9 months ago

I jumped to team red this build.
I have been very happy with my 7900XTX.
4K max settings / FPS on every game I’ve thrown at it.
I don’t play the latest games, so I guess I could hit a wall if I play the recent AAA releases, but many times they simply don’t interest me.

[-] OfficialThunderbolt@beehaw.org 2 points 9 months ago

Apple. Their own processors have both GPUs and AI accelerators. But for some reason, the industry refuses to use them.

this post was submitted on 08 Sep 2023
315 points (100.0% liked)

Technology

37208 readers
224 users here now

Rumors, happenings, and innovations in the technology sphere. If it's technological news or discussion of technology, it probably belongs here.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS