this post was submitted on 16 Sep 2024
77 points (92.3% liked)

PC Gaming

8250 readers
535 users here now

For PC gaming news and discussion. PCGamingWiki

Rules:

  1. Be Respectful.
  2. No Spam or Porn.
  3. No Advertising.
  4. No Memes.
  5. No Tech Support.
  6. No questions about buying/building computers.
  7. No game suggestions, friend requests, surveys, or begging.
  8. No Let's Plays, streams, highlight reels/montages, random videos or shorts.
  9. No off-topic posts/comments.
  10. Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] tunetardis@lemmy.ca 10 points 3 days ago (2 children)

They’re going to keep making more powerful hardware either way, since parallel processing capability supports graphics and AI just fine.

It's not quite as simple as that. AI needs less precision than regular graphics, so chips developed with AI in mind do not necessarily translate into higher performance for other things.

In science/engineering, people want more—not less—precision. So we look for GPUs with capable 64-bit processing, while AI is driving the industry in the other direction, from 32 down to 16.

[–] catloaf@lemm.ee 4 points 3 days ago (1 children)

For science and engineering, workstation cards like the A6000 aren't going anywhere.

[–] henfredemars@infosec.pub 1 points 3 days ago

That’s true, but I would like to see improvements driven along the consumer segment also. AI rendering is a nice software addition but I could easily see it becoming a distraction from hardware improvements.

Consumers generally can’t just throw more money at a problem in the way that professional and business can.

[–] averyminya@beehaw.org 3 points 3 days ago

It's funny because we don't even need GPU's. There's tech that offloads the model's "search" to an analog computer which is ~98% accurate for a fraction of the energy.

I imagine NVIDIA isn't too excited about that side of AI, though.