this post was submitted on 22 Nov 2024
756 points (98.1% liked)

Comic Strips

17912 readers
3887 users here now

Comic Strips is a community for those who love comic stories.

The rules are simple:

Web of links

founded 2 years ago
MODERATORS
756
submitted 7 months ago* (last edited 7 months ago) by Joker@sh.itjust.works to c/comicstrips@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[–] brucethemoose@lemmy.world 1 points 7 months ago* (last edited 7 months ago)

It turns out these clusters are being used very inefficiently, seeing how Qwen 2.5 was trained with a fraction of the GPUs and is clobbering models from much larger clusters.

One could say Facebook, OpenAI, X and such are "hoarding" H100s but are not pressured to utilize them efficiently since they are so GPU unconstrained.

Google is an interesting case, as Gemini is getting better quickly, but they presumably use much more efficient/cheap TPUs to train.