this post was submitted on 13 Jul 2023
31 points (100.0% liked)
Technology
23 readers
2 users here now
This magazine is dedicated to discussions on the latest developments, trends, and innovations in the world of technology. Whether you are a tech enthusiast, a developer, or simply curious about the latest gadgets and software, this is the place for you. Here you can share your knowledge, ask questions, and engage in discussions on topics such as artificial intelligence, robotics, cloud computing, cybersecurity, and more. From the impact of technology on society to the ethical considerations of new technologies, this category covers a wide range of topics related to technology. Join the conversation and let's explore the ever-evolving world of technology together!
founded 2 years ago
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
LOL, I love kbin's public downvote records. I quoted a bunch of different sources demonstrating that you're wrong, and rather than own up to it and apologize for preaching from atop Mt. Dunning-Kruger, you downvoted me and ran off.
I advise you to step out of whatever echo chamber you've holed yourself up in and learn a bit about AI before opining on it further.
My last response didn’t post for some reason. The mistake you’re making is that a neural network is not a neural simulation. It’s relatively simple math, just on a very large scale. I think you mentioned earlier, for example, you played with PyTorch. You should then know that NN stack is based on vector math. You’re making assumptions based on terminology but when you read deeper you’ll see what I mean.
I said it was a neural network.
You said it wasn't.
I asked you for a link.
You told me to do your homework for you.
I did your homework. Your homework says it's a neural network. I suggest you read it, since I took the time to find it for you.
Anyone who knows the first thing about neural networks knows that, yes, artificial neurons are simulated with matrix multiplications, why is why people use GPUs to do them. The simulations are not down to the molecule becuase they don't need to be. The individual neurons are relatively simple math, but when you get into billions of something, you don't need extreme complexity for new properties to emerge (in fact, the whole idea of emergent properties is that they arise from collections of simple things, like the rules of the Game of Life, for instance, which are far simpler than simulated neurons). Nothing about this makes me wrong about what I'm talking about for the purposes of copyright. Neural networks store concepts. They don't archive copies of data.