this post was submitted on 19 Nov 2024
891 points (97.7% liked)

People Twitter

5236 readers
1900 users here now

People tweeting stuff. We allow tweets from anyone.

RULES:

  1. Mark NSFW content.
  2. No doxxing people.
  3. Must be a tweet or similar
  4. No bullying or international politcs
  5. Be excellent to each other.

founded 1 year ago
MODERATORS
 
(page 2) 50 comments
sorted by: hot top controversial new old
[–] Nurse_Robot@lemmy.world 18 points 15 hours ago (6 children)

sigh people do talk about this, they complain about it non-stop. These same people probably aren't using it as intended, or are deliberately trying to farm a "gotcha" response. AI is a very neat tool which can do a lot of things well, but it's important to recognize its limitations. I don't use it for things I don't understand because I won't recognize if it's spitting out nonsense, but for topics I do understand it's hard to overstate how efficient and time saving it is.

[–] zarkanian@sh.itjust.works 6 points 12 hours ago

"Give me a vegan recipe using " has been flawless. The recipes are decent, although they tend to use the same spices over and over.

[–] Paradigm_shift@sh.itjust.works 4 points 12 hours ago

I sometimes use it to "convert" preexisting bulletpoints or informal notes into a professional sounding business email. I already know all the information so proofreading the final product doesn't take a lot of time.

I think a lot of people who shit on AI forget that some people struggle with putting their thoughts into words. Especially if they aren't writing in their native language.

[–] Rekorse@sh.itjust.works -1 points 15 hours ago (1 children)

Efficiency depends on the cost doesnt it?

[–] Nurse_Robot@lemmy.world 2 points 14 hours ago (1 children)

The cost to me, the user, is nothing

[–] taladar@sh.itjust.works -3 points 12 hours ago (1 children)

Sorry to hear that you consider your time worthless. Have you tried therapy for that?

[–] Nurse_Robot@lemmy.world 1 points 7 hours ago

There's something so uniquely funny about being too stupid to insult someone properly. Thanks for the chuckle

load more comments (3 replies)
[–] 1stTime4MeInMCU@mander.xyz 74 points 20 hours ago (1 children)

I’m convinced people who can’t tell when a chat bot is hallucinating are also bad at telling whether something else they’re reading is true or not. What online are you reading that you’re not fact checking anyway? If you’re writing a report you don’t pull the first fact you find and call it good, you need to find a couple citations for it. If you’re writing code, you don’t just write the program and assume it’s correct, you test it. It’s just a tool and I think most people are coping because they’re bad at using it

[–] BluesF@lemmy.world 7 points 20 hours ago (1 children)

Yeah. GPT models are in a good place for coding tbh, I use it every day to support my usual practice, it definitely speeds things up. It's particularly good for things like identifying niche python packages & providing example use cases so I don't have to learn shit loads of syntax that I'll never use again.

[–] Aceticon@lemmy.world 30 points 16 hours ago (3 children)

In other words, it's the new version of copying code from Stack Overflow without going to the trouble of properly understanding what it does.

[–] archomrade@midwest.social 4 points 10 hours ago (2 children)

The usefulness of Stack Overflow or a GPT model completely depends on who is using it and how.

It also depends on who or what is answering the question, and I can't tell you how many times someone new to SO has been scolded or castigated for needing/wanting help understanding something another user thinks is simple. For all of the faults of GPT models, at least they aren't outright abusive to novices trying to learn something new for themselves.

load more comments (2 replies)
[–] BluesF@lemmy.world 3 points 12 hours ago

I know how to write a tree traversal, but I don't need to because there's a python module that does it. This was already the case before LLMs. Now, I hardly ever need to do a tree traversal, honestly, and I don't particularly want to go to the trouble of learning how this particular python module needs me to format the input or whatever for the one time this year I've needed to do one. I'd rather just have something made for me so I can move on to my primary focus, which is not tree traversals. It's not about avoiding understanding, it's about avoiding unnecessary extra work. And I'm not talking about saving the years of work it takes to learn how to code, I'm talking about the 30 minutes of work it would take for me to learn how to use a module I might never use again. If I do, or if there's a problem I'll probably do it properly the second time, but why do it now if there's a tool that can do it for me with minimum fuss?

[–] Rekorse@sh.itjust.works 5 points 15 hours ago

Pft you must have read that wrong, its clearly turning them into master programmer one query at a time.

[–] lightnsfw@reddthat.com 1 points 9 hours ago

I usually tell it "using only information found on applicationwebsite.com " that works pretty well at least to get me in the ballpark to find the answer I'm looking for.

[–] glitchdx@lemmy.world 3 points 12 hours ago

in my use case, the hallucinations are a good thing. I write fiction, in a fictional setting that will probably never actually become a book. If i like what gpt makes up, I might keep it.

Usually, I'll have a conversation going into detail about a subject, this is me explaining the subject to gpt, then having gpt summarize everything it learned about the subject. I then plug that summary into my wiki of lore that nobody will ever see. Then move on to the next subject. Also gpt can identify potential connections between subjects that I didn't think about, and wouldn't have if it didn't hallucinate them.

[–] TrickDacy@lemmy.world 15 points 17 hours ago

Probably because they're not checking them

[–] Randelung@lemmy.world -3 points 7 hours ago (1 children)

Because realistically, that time is zero.

load more comments (1 replies)
[–] Sam_Bass@lemmy.world 6 points 17 hours ago

They're trying not to lose money on the developments

[–] hoshikarakitaridia@lemmy.world 16 points 21 hours ago (3 children)

Because in a lot of applications you can bypass hallucinations.

  • getting sources for something
  • as a jump off point for a topic
  • to get a second opinion
  • to help argue for r against your position on a topic
  • get information in a specific format

In all these applications you can bypass hallucinations because either it's task is non-factual, or it's verifiable while promoting, or because you will be able to verify in any of the superseding tasks.

Just because it makes shit up sometimes doesn't mean it's useless. Like an idiot friend, you can still ask it for opinions or something and it will definitely start you off somewhere helpful.

[–] WalnutLum@lemmy.ml 22 points 18 hours ago (2 children)

All LLMs are text completion engines, no matter what fancy bells they tack on.

If your task is some kind of text completion or repetition of text provided in the prompt context LLMs perform wonderfully.

For everything else you are wading through territory you could probably do easier using other methods.

load more comments (2 replies)
[–] ms_lane@lemmy.world 25 points 20 hours ago (1 children)

Also just searching the web in general.

Google is useless for searching the web today.

[–] fibojoly@sh.itjust.works 1 points 15 hours ago

Not if you want that thing that everyone is on about. Don't you want to be in with the crowd?! /s

[–] ohwhatfollyisman@lemmy.world 3 points 20 hours ago (2 children)

so, basically, even a broken clock is right twice a day?

[–] dev_null@lemmy.ml 5 points 16 hours ago (2 children)

Yes, but for some tasks mistakes don't really matter, like "come up with names for my project that does X". No wrong answers here really, so an LLM is useful.

[–] ohwhatfollyisman@lemmy.world 0 points 12 hours ago (2 children)

great value for all that energy it expends, indeed!

[–] dev_null@lemmy.ml 1 points 9 hours ago

Can't agree

[–] archomrade@midwest.social 1 points 10 hours ago (2 children)

The energy expenditure for GPT models is basically a per-token calculation. Having it generate a list of 3-4 token responses would barely be a blip compared to having it read and respond entire articles.

There might even be a case for certain tasks with a GPT model being more energy efficient than making multiple google searches for the same. Especially considering all the backend activity google tacks on for tracking users and serving ads, complaining about someone using a GPT model for something like generating a list of words is a little like a climate activist yelling at someone for taking their car to the grocery store while standing across the street from a coal-burning power plant.

load more comments (2 replies)
[–] Rekorse@sh.itjust.works -3 points 15 hours ago (1 children)

How is that faster than just picking a random name? Noone picks software based on name.

[–] dev_null@lemmy.ml 4 points 14 hours ago* (last edited 14 hours ago) (2 children)

And yet virtually all of software has names that took some thought, creativity, and/or have some interesting history. Like the domain name of your Lemmy instance. Or Lemmy.

And people working on something generally want to be proud of their project and not name it the first thing that comes to mind, but take some time to decide on a name.

load more comments (2 replies)
[–] onionsinmypores@sh.itjust.works 5 points 17 hours ago* (last edited 17 hours ago)

No, maybe more like, even a functional clock is wrong every 0.8 days.
https://superuser.com/questions/759730/how-much-clock-drift-is-considered-normal-for-a-non-networked-windows-7-pc

The frequency is probably way higher for most LLMs though lol

[–] Snowclone@lemmy.world 13 points 21 hours ago* (last edited 21 hours ago) (2 children)

I only use it for complex searches with results I can usually parse myself like ''list 30 typical household items without descriptions or explainations with no repeating items'' kind of thing.

[–] ohwhatfollyisman@lemmy.world 9 points 20 hours ago

great value for all that energy it expends, indeed!

[–] Varyk@sh.itjust.works -4 points 20 hours ago

it's because everyone stopped using it, right?

at least months ago?

[–] spankmonkey@lemmy.world 14 points 21 hours ago (4 children)

Because most people are too lazy to bother with making sure the results are accurate when they sound plausible. They want to believe the hype, and lack critical thinking.

[–] Chip_Rat@lemmy.world 2 points 14 hours ago (1 children)

I don't want to believe any hype! I just want to be able to ask "hey Chatgtp, I'm looking for a YouTube video by technology connections where he discusses dryer heat pumps." And not have it spit out "it's called "the neat ways your dryer heat pumps save energy!"

And it is not, that video doesn't exist. And it's even harder to disprove it on first glance because the LLM is mimicing what Alex would have called the video. So you look and look with your sisters very inefficient PS4 controller-to-youtube interface... And finally ask it again and it shy flowers you....

But I swear he talked about it ?!?! Anyone?!?

[–] gamermanh@lemmy.dbzer0.com 2 points 13 hours ago* (last edited 8 hours ago) (1 children)

He hasn't

I think in a recent video he mentioned he will soon, but he hasn't done a video with even a segment on heat pumps in dryers yet

Fairly confident in this, recently finished a rewatch of basically all his content

[–] Chip_Rat@lemmy.world 2 points 10 hours ago

Damn it... I was sure he mentioned them briefly in one of his heat pump videos but I trust you over Chatgtp...

He should do a video! I am constantly enchanted by his heat pump explainers... I don't know why but it's one of those concepts that's just a bit out of my wheelhouse. So I always "knew" how it worked. But the lightbulb moment. The aha! Pure crack.

load more comments (3 replies)
[–] callcc@lemmy.world 8 points 20 hours ago

It's usually good for ecosystems with good and loads of docs. Whenever docs are scarce the results become shitty. To me it's mostly a more targeted search engine without the crap (for now)

[–] UnderpantsWeevil@lemmy.world 0 points 10 hours ago (1 children)

In another thread, I was curious about the probability of reaching the age of 60 while living in the US.

Google gave me an assortment of links to people asking similar questions on Quora, and to some generic actuarial data, and to some totally unrelated bullshit.

ChatGPT gave me a multi-paragraph response referencing its data sources and providing both a general life expectancy and a specific answer broken out by gender. I asked ChatGPT how it reached this answer, and it proceeded to show its work. If I wanted to verify the work myself, ChatGPT gave me source material to cross-check and the calculations it used to find the answer. Google didn't even come close to answering the question, much less producing the data it used to reach the answer.

I'm as big an AI skeptic as anyone, but it can't be denied that generic search engines have degraded significantly. I feel like I'm using Alta Vista in the 90s whenever I query Google in the modern day. The AI systems do a marginally better job than old search engines were doing five years ago, before enshittification hit with full force.

It sucks that AI is better, but it IS better.

load more comments (1 replies)
[–] Eheran@lemmy.world 4 points 20 hours ago

Remember when you had to have extremely niche knowledge of "banks" in a microcontroller to be able to use PWM on 2 pins with different frequencies?

Yes, I remember what a pile of shit it was to try and find out why xyz is not working while x and y and z work on their own. GPT usually gets me there after some tries. Not to mention how much faster most of the code is there, from A to Z, with only little to tweak to get it where I want (since I do not want to be hyper specific and/or it gets those details wrong anyway, as would a human without massive context).

[–] RedditWanderer@lemmy.world 5 points 21 hours ago* (last edited 21 hours ago)

Big businesses know, they even ask people like me to add extra measures in place. I like to call it the concorde effect. Youre trying to make a plane that can shove air out of the way faster than it wants to move, and this takes an enormous amount of energy that isn't worth the time save, or the cost. Even if you have higher airspeed when it works, if your plane doesn't make it to destination it isn't "faster".

We hear a lot about the downsides of AI, except that doesn't fit the big corpo narrative and people don't care enough really. If youre just a consumer who has no idea how this really works, the investments companiess make into shoving it everywhere makes it seem like it's not a problem and it looks like there's only AI hype and no party poopers.

[–] damnthefilibuster@lemmy.world 3 points 20 hours ago

What are you talking about? I don’t verify anything that ChatGPT gives me.

[–] pixxelkick@lemmy.world 4 points 21 hours ago

Gippity is pretty good at getting me 90% of the way there.

It usually sets me up with at least all the terms and etc I now know to google, whereas before I wouldnt even know what I am looking for in the first place.

Also not gonna lie, search engines are even worse than gippity for accuracy often.

And Ive had to fight with so many cases of garbage documentation lately that gippity genuinely does the job better, because it has all the random comments from issues and solutions in its data.

Usually once I have my sort of key terms I need to dig into, I can use youtube/google and get more specific information though, and thats the last 10%

[–] PixelatedSaturn@lemmy.world 2 points 20 hours ago

You have to understand it well enough to know what stuff you can rely on. On the other hand nowadays there are often sources there, so it's easy to check.

load more comments
view more: ‹ prev next ›