this post was submitted on 02 May 2024
131 points (98.5% liked)

Games

16723 readers
609 users here now

Video game news oriented community. No NanoUFO is not a bot :)

Posts.

  1. News oriented content (general reviews, previews or retrospectives allowed).
  2. Broad discussion posts (preferably not only about a specific game).
  3. No humor/memes etc..
  4. No affiliate links
  5. No advertising.
  6. No clickbait, editorialized, sensational titles. State the game in question in the title. No all caps.
  7. No self promotion.
  8. No duplicate posts, newer post will be deleted unless there is more discussion in one of the posts.
  9. No politics.

Comments.

  1. No personal attacks.
  2. Obey instance rules.
  3. No low effort comments(one or two words, emoji etc..)
  4. Please use spoiler tags for spoilers.

My goal is just to have a community where people can go and see what new game news is out for the day and comment on it.

Other communities:

Beehaw.org gaming

Lemmy.ml gaming

lemmy.ca pcgaming

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] aard@kyu.de 5 points 6 months ago (1 children)

For AI and compute… They’re far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins

I got a W6800 recently. I know a nvidia model of the same generation would be faster for AI - but that thing is fast enough to run stable diffusion variants with high resolution pictures locally without getting too annoyed.

[–] crispyflagstones@sh.itjust.works 4 points 6 months ago* (last edited 6 months ago) (2 children)

The completely different software stack is a killer. It's not that you can't find versions of a model to run, but almost everything that hits the GPU for compute is going to be targeting CUDA, not RocM. From a compatibility standpoint alone this killed AMD for me. I just do not want to spend my time fighting the stack to get these models running.

[–] WormFood@lemmy.world 4 points 6 months ago (1 children)

on the one hand, cuda is vendor lock-in and if we'd all just agreed on an open standard decades ago then we wouldn't be in this mess

but on the other hand, rocm is crap and adaptivecpp is very half baked right now, at least in my limited experience

[–] crispyflagstones@sh.itjust.works 2 points 6 months ago* (last edited 6 months ago)

Yeah, it's not that I like this state of affairs, but right now the vendor lock-in is so one-sided that it's hard to say there's a viable alternative to CUDA. I hope that changes one day.

[–] aard@kyu.de 1 points 6 months ago

Admittedly I'm just toying around for entertainment purposes - but I didn't really have any problems of getting anything I wanted to try out with rocm support. Bigger annoyance was different projects targetting specific distributions or specific software versions (mostly ancient python), but as I'm doing everything in containers anyway that also was manageable.