this post was submitted on 04 Sep 2024
595 points (94.5% liked)

Technology

59038 readers
3862 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Artificial intelligence is worse than humans in every way at summarising documents and might actually create additional work for people, a government trial of the technology has found.

Amazon conducted the test earlier this year for Australia’s corporate regulator the Securities and Investments Commission (ASIC) using submissions made to an inquiry. The outcome of the trial was revealed in an answer to a questions on notice at the Senate select committee on adopting artificial intelligence.

The test involved testing generative AI models before selecting one to ingest five submissions from a parliamentary inquiry into audit and consultancy firms. The most promising model, Meta’s open source model Llama2-70B, was prompted to summarise the submissions with a focus on ASIC mentions, recommendations, references to more regulation, and to include the page references and context.

Ten ASIC staff, of varying levels of seniority, were also given the same task with similar prompts. Then, a group of reviewers blindly assessed the summaries produced by both humans and AI for coherency, length, ASIC references, regulation references and for identifying recommendations. They were unaware that this exercise involved AI at all.

These reviewers overwhelmingly found that the human summaries beat out their AI competitors on every criteria and on every submission, scoring an 81% on an internal rubric compared with the machine’s 47%.

you are viewing a single comment's thread
view the rest of the comments
[–] Darkassassin07@lemmy.ca 111 points 1 month ago (13 children)

And for a much much smaller paycheck.

All corporate gives af about.

[–] pennomi@lemmy.world 59 points 1 month ago (12 children)

It might be all I care about. Humans might always be better, but AI only has to be good enough at something to be valuable.

For example, summarizing an article might be incredibly low stakes (I’m feeling a bit curious today), or incredibly high stakes (I’m preparing a legal defense), depending on the context. An AI is sufficient for one use but not the other.

[–] Grandwolf319@sh.itjust.works 2 points 1 month ago* (last edited 1 month ago) (6 children)

I mean, what you’re essentially implying is, what if we could do a lot of things that we do today, but faster and less quality.

Imo we have too much things today and very few are worth their salt, so this is the opposite of the right direction.

[–] pennomi@lemmy.world 1 points 1 month ago (1 children)

That’s not what I’m implying. What I’m saying is that wasting time and effort on quality is pointless when the threshold for success is low.

For example, I could use aerospace quality parts (perfectly machined to micron-level tolerances) to build a toaster. However, while this would not increase the performance meaningfully, the cost would be orders of magnitude greater. Instead I can use shitty off-the-shelf parts because it doesn’t really make a difference.

Maybe in other words, engineering tolerances apply to LLMs too. They’re crude devices, but it’s totally fine if you have a crude problem.

[–] Grandwolf319@sh.itjust.works 2 points 1 month ago (2 children)

That’s not what I’m implying. What I’m saying is that wasting time and effort on quality is pointless when the threshold for success is low.

Yes and my response to that is for some people maybe, for others they don’t want a low threshold, they want few good articles instead of spam of low quality.

Maybe in other words, engineering tolerances apply to LLMs too. They’re crude devices, but it’s totally fine if you have a crude problem.

Exactly, I’m saying there is no objective crude problem. You might be okay with simple summaries but I want every single piece of information I consume to have a very high bar.

[–] AA5B@lemmy.world 2 points 1 month ago (1 children)

What if you’re reading Lemmy, and you don’t really feel like reading the article. Is the headline likely to tell you all you need to know or is the ai summary likely to find more info and without the clickbait?

[–] Grandwolf319@sh.itjust.works 1 points 1 month ago* (last edited 1 month ago)

Imo it’s on me to either read the article or be okay with not being informed. Don’t get me wrong, a summery is good, but not when it’s not reliable and the article is a click away, some might have a different comfort level.

[–] pennomi@lemmy.world 1 points 1 month ago (1 children)

Sure, go for it. But good luck paying an army of copywriters to summarize every article you read.

[–] Grandwolf319@sh.itjust.works 1 points 1 month ago

No summery is better than a bad summery, it would encourage you to actually read the source.

load more comments (4 replies)
load more comments (9 replies)
load more comments (9 replies)