this post was submitted on 03 Dec 2024
148 points (96.8% liked)

Technology

59736 readers
2801 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
 

If even half of Intel's claims are true, this could be a big shake up in the midrange market that has been entirely abandoned by both Nvidia and AMD.

top 50 comments
sorted by: hot top controversial new old
[–] Blackmist@feddit.uk 1 points 17 minutes ago (1 children)

How is compatibility with older games now?

Because I'm not buying a GPU unless it works with everything.

[–] empireOfLove2@lemmy.dbzer0.com 2 points 8 minutes ago

We'll see come launch, but even the original Arc cards work totally fine with basically all DX9 games now. Arc fell victim to half baked drivers because Intel frankly didn't know what they were doing. That's a few years behind them now.

Intel designed their uarch to be DX 11/12/Vulkan based and not support hardware level DX9 and older drawcalls, which is a reasonable choice for a ground-up implementation- however it does also mean that it only runs older graphics interpreters using a translation/emulation layer, turning DX9 into DX12. And driver emulation is an always imperfect science.

[–] Assman@sh.itjust.works 38 points 11 hours ago (4 children)

All these weird numeric names. I'm gonna build a GPU and name it Jonathan.

[–] ms_lane@lemmy.world 1 points 50 minutes ago

Hello, My name is Roger!

[–] jonathan@lemmy.zip 10 points 7 hours ago
[–] lime@feddit.nu 8 points 9 hours ago

sorry, apple already took that one. call it Jeff or something.

[–] Brickhead92@lemmy.world 9 points 11 hours ago

Just don't name it Steve. You're in for a world of troubles with GPU Steve.

[–] brucethemoose@lemmy.world 30 points 13 hours ago* (last edited 13 hours ago) (2 children)

If they double up the VRAM with a 24GB card, this would be great for a "self hosted LLM" home server.

3060, 3090 prices have been rising like crazy because Nvidia is vram gouging and AMD inexplicably refuses to compete. Even ancient P40s (double vram 1080 TIs with no display) are getting expensive. 16GB on the A770 is kinda meager, but 24GB is the point where you can fit the Qwen 2.5 32B models that are starting to perform like the big corporate API ones.

And if they could fit 48GB with new ICs... Well, it would sell like mad.

[–] Psythik@lemmy.world 10 points 10 hours ago (4 children)

I always wondered who they were making those mid- and low-end cards with a ridiculous amount of VRAM for... It was you.

All this time I thought they were scam cards to fool people who believe that bigger number always = better.

[–] sugar_in_your_tea@sh.itjust.works 7 points 10 hours ago (1 children)

Yeah, AMD and Intel should be running high VRAM SKUs for hobbyists. I doubt it'll cost them that much to double the RAM, and they could mark them up a bit.

I'd buy the B580 if it had 24GB RAM, at 12GB, I'll probably give it a pass because my 6650 XT is still fine.

[–] M600@lemmy.world 2 points 9 hours ago (1 children)

Don’t you need nvidia cards to run ai stuff?

[–] sugar_in_your_tea@sh.itjust.works 7 points 9 hours ago* (last edited 9 hours ago)

Nah, ollama works w/ AMD just fine, just need a model w/ enough VRAM.

I'm guessing someone would get Intel to work as well if they had enough VRAM.

[–] brucethemoose@lemmy.world 6 points 9 hours ago

Also "ridiculously" is relative lol.

The Llm/workstation crowd would buy a 48GB 4060 without even blinking, if that were possible. These workloads are basically completely vram constrained.

[–] brucethemoose@lemmy.world 1 points 9 hours ago (1 children)

Like the 3060? And 4060 TI?

Its ostensibly because they’re “too powerful” for their vram to be cut in half (so 6GB on the 3060 and 8GB on the 4060 TI), but yes, more generally speaking these are sweetspot for vram heavy workstation/compute workloads. Local LLMs are just the most recent one.

Nvidia cuts vram at the high end to protect their server/workstation cards, AMD does it… Just because?

[–] Psythik@lemmy.world 1 points 7 hours ago* (last edited 7 hours ago) (1 children)

More like back in the day when you would see vendors slapping 1GB on a card like the Radeon 9500, when the 9800 came with 128MB.

[–] brucethemoose@lemmy.world 2 points 6 hours ago* (last edited 6 hours ago)

Ah yeah those were the good old days when vendors were free to do that, before AMD/Nvidia restricted them. It wasn't even that long ago, I remember some AMD 7970s being double VRAM.

And, again, I'd like to point out how insane this restriction is for AMD given their market struggles...

[–] Fedegenerate@lemmynsfw.com 2 points 9 hours ago* (last edited 8 hours ago) (2 children)

An LLM card with quicksync would be the kick I need to turn my n100 mini into a router. Right now, my only drive to move is that my storage is connected via usb. SATA is just not enough value for a whole new box. £300 for Ollama, much faster ml in immich etc and all the the transcodes I could want would be a "buy now figure the rest out later" moment.

[–] brucethemoose@lemmy.world 2 points 8 hours ago (1 children)

Oh also you might look at Strix Halo from AMD in 2025?

Its IGP is beefy enough for LLMs, and it will be WAY lower power than any dGPU setup, with enough vram to be "sloppy" and run stuff in parallel with a good LLM.

[–] Fedegenerate@lemmynsfw.com 2 points 8 hours ago

*adds to wishlist

[–] brucethemoose@lemmy.world 2 points 8 hours ago (1 children)

You could get that with 2x B580s in a single server I guess, though yoi could have already done that with the A770s.

[–] Fedegenerate@lemmynsfw.com 3 points 8 hours ago (1 children)

... That's nuts. I only just graduated to a mini from a pi, I didnt consider a dual GPU setup. Arbitrary budget aside, I should have added an "idle power" constraint too. Reasonable to assume that as soon as LLMs get involved all concept of "power efficient" goes out the window. Don't mind me, just wishing for a unicorn.

[–] brucethemoose@lemmy.world 2 points 8 hours ago

Strix Halo is your unicorn, idle power should be very low (assuming AMD VCE is OK over quicksync)

[–] vzq@lemmy.world 40 points 16 hours ago (2 children)

Intel GPU claims are NEVER true.

[–] MudMan@fedia.io 24 points 15 hours ago (1 children)

Meh, I ended up with an A770 for a repurposed PC and it's been pretty solid, especially for the discounted price I got. I get that there were some driver growing pains, but I'm not in a hurry to replace that thing, it was a solid gamble.

[–] brucethemoose@lemmy.world 7 points 13 hours ago* (last edited 13 hours ago) (5 children)

The A770 was definitely a "fine wine" card from the start. Its raw silicon specs were way stronger than the competition, it just needed to grow into it.

This ones a bit smaller though...

load more comments (5 replies)
[–] unexposedhazard@discuss.tchncs.de 2 points 10 hours ago (1 children)

It hasnt been like that anymore for a while now.

[–] vzq@lemmy.world -1 points 2 hours ago (1 children)

If I had a dime for every time I heard that exact line.

At a certain point it’s a fool me once, fool me twice, fool me fourteen times kinda thing.

[–] ms_lane@lemmy.world 3 points 49 minutes ago

Buy nVidia forever then.

Enjoy the prices!

[–] Buffalox@lemmy.world 12 points 13 hours ago* (last edited 13 hours ago) (2 children)

Funny the Radeon RX 480 came out in 2016 at a similar price. Is that a coincidence?
Incidentally the last great generation offering a midrange GPU at a midrange price. The Nvidia 1060 was great too, and the 1080 is claimed to maybe be one of the best offers of all time. Since then everything has been overpriced.

The RX 480 was later replaced by the 580 which was a slight upgrade at great value too. But then the crypto wave hit, and soon a measly 580 cost nearly $1000!!! Things have never returned quite back to normal since. Too little competition with only Nvidia and AMD.

30 series started to look like a return to good priced cards. Then crypto hit and ruined that. Now we have AI to keep that gravy train going. Once the AI hype dies down maybe we'll see cards return to sane pricing.

load more comments (1 replies)
[–] Juice260@lemmy.world 18 points 16 hours ago (8 children)

I’m reserving judgement of course to see how things actually play out but I do want to throw a cheapest pc together for my nephew and that card would make a fine centerpiece.

load more comments (8 replies)
[–] DarkThoughts@fedia.io 7 points 14 hours ago (3 children)

As someone with a 6650 XT, which is a little slower than the 6700 or 4060, I doubt the increased vram, which is of course still nice, is enough to push it for 1440p. I struggle even in 1080p in some games, but I guess if you're okay with ~40 FPS then you could go that high.

Unfortunately, if the 4060 is roughly the target here, that's still far below what I'm interested in, which is more the upper midrange stuff (and I'd love one with 16 GB vram at least).

At least the price is much more attractive now.

load more comments (3 replies)
[–] sugar_in_your_tea@sh.itjust.works 3 points 13 hours ago (9 children)

Seems like a decent card, but here are my issues:

  • 12 GB RAM - not quite enough to play w/ LLMs
  • a little better than my 6650 XT, but not amazingly so
  • $250 - a little better than RX 7600 and RTX 4060 I guess? Probably?

If it offered more RAM (16GB or ideally 24GB) and stayed under $300, I'd be very interested because it opens up LLMs for me. Or if it had a bit better performance than my current GPU, and again stayed under $300 (any meaningful step-up is $350+ from AMD or Nvidia).

But this is just another low to mid-range card, so I guess it would be interesting for new PC builds, but not really an interesting upgrade option. So, pretty big meh to me. I guess I'll check out independent benchmarks in case there's something there. I am considering building a PC for my kids using old parts, so I might get this instead of reusing my old GTX 960, the board I'd use only has PCIe 3.0, so I worry performance would suffer and the GTX 960 may be a better stop-gap.

[–] bluemellophone@lemmy.world 2 points 5 hours ago

There are some smaller Ollama Llama 3.2 models that would fit on 12GB. I’ve run some of the smaller Llama 3.1 models under 10GB on NVIDIA GPUs

[–] SharkAttak@kbin.melroy.org 9 points 9 hours ago (1 children)

12 GB RAM - not quite enough to play w/ LLMs

Good. Not every card has to be about AI, there's enough of those already; we need gaming cards.

Sure, I'm just saying what I would need for this card to be interesting. It's not much better than what I have, and the extra VRAM isn't enough to significantly change what I can do with it.

So it either needs to have better performance or more VRAM for me to be interested.

It's a decent choice for new builds, but I don't really think it makes sense as an upgrade for pretty much any card that exists today.

[–] exu@feditown.com 3 points 11 hours ago (1 children)

If their claims are true, I'd say this is a decent upgrade from my RX 6600 XT and I'm very likely buying one.

Sounds like a ~10% upgrade, but I'd definitely wait for independent reviews because that could be optimistic. It's certainly priced about even with the 6600 XT.

But honestly, if you can afford an extra $100 or so, you'd be better off getting a 6800 XT. It has more VRAM and quite a bit better performance, so it should last you a bit longer.

load more comments (6 replies)
load more comments
view more: next ›