this post was submitted on 31 Jul 2024
284 points (96.4% liked)

Technology

59414 readers
3554 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Meta "programmed it to simply not answer questions," but it did anyway.

top 50 comments
sorted by: hot top controversial new old
[–] snooggums@midwest.social 123 points 3 months ago (2 children)

Hallucinating is a fancy term for BEING WRONG.

Unreliable bullshit generator is still unreliable. Imagine that!

[–] doodledup@lemmy.world 51 points 3 months ago* (last edited 3 months ago) (2 children)

AI doesn't know what's wrong or correct. It hallucinates every answer. It's up to the supervisor to determine whether it's wrong or correct.

Mathematically verifying the correctness of these algorithms is a hard problem. It's intentional and the trade-off for the incredible efficiency.

Besides, it can only "know" what it has been trained on. It shouldn't be suprising that it cannot answer about the Trump shooting. Anyone who thinks otherwise simply doesn't know how to use these models.

[–] snooggums@midwest.social 13 points 3 months ago (4 children)

It is impossible to mathematically determine if something is correct. Literally impossible.

At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn't the same thing is consensus, because AI is not intelligent.

If the 'supervisor' has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

[–] doodledup@lemmy.world 2 points 3 months ago* (last edited 3 months ago) (10 children)

It is impossible to mathematically determine if something is correct. Literally impossible.

No, you're wrong. You can indeed prove the correctness of a neural network. You can also prove the correctness of many things. It's the most integral part of mathematics and computer-science.

For example a very simple proof: with the conjecture that an even number is 2k of a number k, then you can prove that the addition of two even numbers is again an even number (and that prove is definite): 2a+2b=2(a+b), since a+b=k for some k.

Obviously, proving more complex mathematical problems like AI is more involved. But that's why we have scientists that work on that.

At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn't the same thing is consensus, because AI is not intelligent.

That is correct. But it's not a limitation. It's by design. It's the tradeoff for the efficiency of the models. It's like lossy JPG compression. You accept some artifacts but in return you get much smaller images and much faster loading times.

But there are indeed "AI"s and neural networks that have been proven correct. This is mostly applied to safety critical applications like airplane collision avoidance systems or DAS. But a language model is not safety critical; so we take full advantage.

If the 'supervisor' has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

You're completely misunderstanding the whole thing. The only reason why it's so incredibly good in many applications is because it's bad in others. It's intentionally designed that way. There are exact algorithms and there approximation algorithms. The latter tend to be much more efficient and usable in practice.

[–] Cethin@lemmy.zip 11 points 3 months ago* (last edited 3 months ago) (20 children)

You can prove some things are correct, like math problems (assuming the axioms they are based on are also correct).

You can't prove that things like events having happened are correct. That's even a philosophical issue with human memory. We can't prove anything in the past actually happened. We can hope that our memory of events is accurate and reliable and work from there, but it can't actually be proven. In theory everything before could have just been implanted into our minds. This is incredibly unlikely (as well as not useful at best), but it can't be ruled out.

If we could prove events in the past are true we wouldn't have so many pseudo-historians making up crazy things about the pyramids, or whatever else. We can collect evidence and make inferences, but we can't prove it because it is no longer happening. There's a chance that we miss something or some information can't be recovered.

LLMs are algorithms that use large amounts of data to identify correlations. You can tune them to give more unique answers or more consistent answers (and other conditions) but they aren't intelligent. They are, at best, correlation finders. If you give it bad data (internet conversations) or incomplete data then it at best will (usually confidently) give back bad information. People who don't understand how they work assume they're actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

[–] rottingleaf@lemmy.world 3 points 3 months ago

You can’t prove that things like events having happened are correct.

You can't so solidly that this shouldn't even be discussed.

What should be is whether you can make a machine capable of reasoning.

There's symbolic logic, so you can maybe some day make a machine that makes correct syllogisms, detects incorrect syllogisms and such.

People who don’t understand how they work assume they’re actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

Sadly there's that archetype of "the narrow-minded not cool scientist against the cool brave inventor" which means that actively dispelling that may do harm. People who don't understand will match the situation with that archetype and it will reinforce their belief.

load more comments (19 replies)
[–] jaybone@lemmy.world 6 points 3 months ago (1 children)

Your proof example is a proof from your discrete structures class. That’s very different than “proving” something like “the Trump assassination attempt was a conspiracy.”

Otherwise we could have gotten rid of courts a long time ago.

[–] doodledup@lemmy.world 0 points 3 months ago* (last edited 3 months ago)

Well obviously. But that was not at all what I said or claimed. I just said that you can prove certain properties of neural networks because others said that you can't. And others also misunderstood LLMs in general. They believe it's an information retrival service, which is wrong.

Besides, your argument, as you've written it, applies to everything. Literally. From Wikipedia, to News, even up to your eyesight. What can you actually prove? I don't understand the point you're making and how that is related to LLMs.

[–] markon@lemmy.world 0 points 3 months ago

Just like us. Sometimes it's better to have bullshit predictions than none.

load more comments (7 replies)
[–] superminerJG@lemmy.world 0 points 3 months ago (1 children)

That is, unless you define correct in mathematical terms. Which no one has done yet.

load more comments (1 replies)
load more comments (2 replies)
[–] markon@lemmy.world 0 points 3 months ago (1 children)

Uhm. Have you ever talked to a human being.

[–] doodledup@lemmy.world 1 points 3 months ago* (last edited 3 months ago)

Human beings are not infallible either.

[–] theterrasque@infosec.pub 1 points 3 months ago (2 children)

That's like saying car crash is just a fancy word for accident, or cat is just a fancy term for animal.

Hallucination is a technical term for this type of AI, and it's inherent to how it works at it's core.

And now I'll let you get back to your hating.

[–] CileTheSane@lemmy.ca 9 points 3 months ago

Hallucination is also wildly misleading. The AI does not believe something that isn't real, it was incorrect in the words it guessed would be appropriate.

load more comments (1 replies)
[–] catloaf@lemm.ee 55 points 3 months ago (4 children)

Kaplan noted that AI chatbots "are not always reliable when it comes to breaking news or returning information in real time," because "the responses generated by large language models that power these chatbots are based on the data on which they were trained, which can at times understandably create some issues when AI is asked about rapidly developing real-time topics that occur after they were trained."

If you're expecting a glorified autocomplete to know about things it doesn't have in its training data, you're an idiot.

[–] Catoblepas@lemmy.blahaj.zone 38 points 3 months ago (1 children)

There are definitely idiots, but these idiots don’t get their ideas of how the world works out of thin air. These AI chatbot companies push the cartoon reality that this is a smart robot that knows things hard in their advertisements, and to learn otherwise you have to either listen to smart people or read a lot of text.

[–] vaultdweller013@sh.itjust.works 1 points 3 months ago

I just assumed that its bs at first, but I also once nearly went unga bunga caveman against a computer from 1978. So I probably have a deeper understanding of how dumb computers can be.

[–] Empricorn@feddit.nl 3 points 3 months ago (10 children)

Yeah, the average person is the idiot here, for something they never asked for, and for something they see no value in. Companies threw billions of dollars at this emerging technology. Many things like Google Search have hallucinating, error-prone AI forced into the main product that is impossible to opt-out or use the (working) legacy version now...

load more comments (10 replies)
[–] brucethemoose@lemmy.world 3 points 3 months ago

Some services will use glorified RAG to put more current info in the context.

But yeah, if it's just the raw model, I'm not sure what they were expecting.

[–] Telorand@reddthat.com 1 points 3 months ago

Sir, are you telling me AI isn't a panacea for conveying facts? /s

[–] randon31415@lemmy.world 47 points 3 months ago* (last edited 3 months ago) (1 children)

The shooting happened after the end of the training date. Like asking windows 95 clippy about 9/11 and it saying it didn't happen.

[–] Evotech@lemmy.world 43 points 3 months ago

Clippy being a 9/11 conspiracy theorist is now canon

[–] expatriado@lemmy.world 12 points 3 months ago

maybe Meta AI is into something

[–] nondescripthandle@lemmy.dbzer0.com 7 points 3 months ago (1 children)

Is it wrong to root this on simply because I hate that shitbag?

[–] intensely_human@lemm.ee 4 points 3 months ago (1 children)

Hatred is a path to the dark side.

As evidenced by you now rooting for misinformation.

[–] nondescripthandle@lemmy.dbzer0.com 1 points 3 months ago* (last edited 3 months ago)

Oh I'm far to pragmatic to believe that. If truth isn't working, then what choice do you really have?

[–] InternetUser2012@lemmy.today 5 points 3 months ago

Well, if the chatbot learned anything from Dementia Don the racist rapist with 34 felonies that can't complete a coherent sentence, it learned that you never tell the truth.

[–] Papanca@lemmy.world 4 points 3 months ago

Does this AI work with real time info?

load more comments
view more: next ›