this post was submitted on 15 Nov 2024
63 points (100.0% liked)

Futurology

1776 readers
194 users here now

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] hendrik@palaver.p3x.de 7 points 4 hours ago (1 children)

What kind of AI workloads are these NPUs good at? I mean it can't be most of generative AI like LLMs, since that's mainly limited by the memory bandwith and at this point it doesn't really matter if you have a NPU, GPU or CPU... You first need lots of fast RAM and a wide interface to it.

[–] SlopppyEngineer@lemmy.world 1 points 1 hour ago (1 children)

That's why NPU will have high bandwidth memory on chip. They're also low precision to save power but massively parallel. A GPU and CPU can do it too, but less optimized.

[–] hendrik@palaver.p3x.de 1 points 39 minutes ago* (last edited 36 minutes ago)

That was my question... How much on-chip memory do they have? And what are applications for that amount of memory? I think an image generator needs like 4-5GB and a LLM that's smart enough as a general porpose chatbot needs like 8-10GB. More will be better. And at that point you'd better make it unified memory like with the M-series Macs or other APUs? Or this isn't targeted at generative AI but some other applications. Hence my question.