this post was submitted on 16 Sep 2024
77 points (92.3% liked)

PC Gaming

8250 readers
535 users here now

For PC gaming news and discussion. PCGamingWiki

Rules:

  1. Be Respectful.
  2. No Spam or Porn.
  3. No Advertising.
  4. No Memes.
  5. No Tech Support.
  6. No questions about buying/building computers.
  7. No game suggestions, friend requests, surveys, or begging.
  8. No Let's Plays, streams, highlight reels/montages, random videos or shorts.
  9. No off-topic posts/comments.
  10. Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Ghostalmedia@lemmy.world 5 points 3 days ago (1 children)

Thanks! So, from what I grok, the claim is basically that the games could probably run fine if they were written and optimized properly, but since they’re probably not, people have to buy a GPU that applies a bandaid solution. Right?

[–] warm@kbin.earth 8 points 3 days ago

Yep. As more people buy GPUs that have the capabilities to use machine learning upscaling (the bandaid) then the more likely developers are to use it instead of spending time improving performance.

I see it the most in Unreal Engine games, Unreal Engine allows devs to make a "realistic" style game fast, but performance is often left in the dirt. UE also has some of the worst anti-aliasing out of the box, so DLSS for example, is a good catch all to try and improve framerates and provide some AA, but instead you just get a lot of blur and poor graphical fidelity. The issues probably don't exist at higher resolutions, like 4K (which is maybe what they develop with), but the majority of people still use 1080p.

Oops sorry for the rant! I just got pissed off with it again recently in Satisfactory!