this post was submitted on 22 Dec 2024
521 points (96.1% liked)

Technology

60103 readers
2113 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Eldritch@lemmy.world 23 points 4 days ago (3 children)

Computers have always been good at pattern recognition. This isn't new. LLM are not a type of actual AI. They are programs capable of recognizing patterns and Loosely reproducing them in semi randomized ways. The reason these so-called generative AI Solutions have trouble generating the right number of fingers. Is not only because they have no idea how many fingers a person is supposed to have. They have no idea what a finger is.

The same goes for code completion. They will just generate something that fills the pattern they're told to look for. It doesn't matter if it's right or wrong. Because they have no concept of what is right or wrong Beyond fitting the pattern. Not to mention that we've had code completion software for over a decade at this point. Llms do it less efficiently and less reliably. The only upside of them is that sometimes they can recognize and suggest a pattern that those programming the other coding helpers might have missed. Outside of that. Such as generating act like whole blocks of code or even entire programs. You can't even get an llm to reliably spit out a hello world program.

[–] makuus@pawb.social 5 points 4 days ago (2 children)

I never know what to think when I come across a comment like this one—which does describe, even if only at a surface level, how an LLM works—with 50% downvotes. Like, are people angry at reality, is that it?

[–] Eldritch@lemmy.world 15 points 4 days ago (1 children)

With as much misinformation that's being spread about regarding LLMs. It would only lose more people's comprehension to go into anything more than a generalization.

The problem is people are being sold AGI. But chat GPT and all these other tools don't even remotely qualify for that. They're really nothing more than a glorified Alice chatbot system on steroids. The one neat new trick to all this is that they've automated the training a bit. But these llms have no more comprehension of their output or the input they were given than something like the old Alice chatbot.

These tools have been described as artificial intelligence to layman for decades at this point. It makes it really hard to change that calcified opinion. People would rather believe that it's some magical thing not just probability and maths.

[–] spankmonkey@lemmy.world 8 points 4 days ago

They are bullshit machines, trained to output something that users think is the right output.

[–] Naz@sh.itjust.works 2 points 4 days ago (1 children)

Downvoting someone on the Internet is easier than tangentially modifying reality in a measurable way

[–] rezifon@lemmy.world 1 points 12 hours ago

Downvoting sounds like a task that’s ripe for automation with AI!

[–] JohnEdwa@sopuli.xyz 2 points 4 days ago (3 children)

"It's part of the history of the field of artificial intelligence that every time somebody figured out how to make a computer do something—play good checkers, solve simple but relatively informal problems—there was a chorus of critics to say, 'that's not thinking'"
-Pamela McCorduck

"AI is whatever hasn't been done yet."
- Larry Tesler

That's the curse of the AI Effect.
Nothing will ever be "an actual AI" until we cross the barrier to an actual human-like general artificial intelligence like Cortana from Halo, and even then people will claim it isn't actually intelligent.

[–] ssfckdt@lemmy.blahaj.zone 8 points 4 days ago (1 children)

I mean, I think intelligence requires the ability to integrate new information into one's knowledge base. LLMs can't do that, they have to be trained on a fixed corpus.

Also, LLMs have a pretty shit-tastic track record of being able to differentiate correct data from bullshit, which is a pretty essential facet of intelligence IMO

[–] JohnEdwa@sopuli.xyz 8 points 4 days ago

LLMs have a perfect track record of doing exactly what they were designed to, take an input and create a plausible output that looks like it was written by a human. They just completely lack the part in the middle that properly understands what it gets as the input and makes sure the output is factually correct, because if it did have that then it wouldn't be an LLM any more, it would be an AGI.
The "artificial" in AI does also stand for the meaning of "fake" - something that looks and feels like it is intelligent, but actually isn't.

[–] Eldritch@lemmy.world 8 points 4 days ago (1 children)

Well at least until those who study intelligence and self-awareness actually come up with a comprehensive definition for it. Something we don't even have currently. Which makes the situation even more silly. The people selling LLMs and AGNs as artificial intelligence are the PT Barnum of the modern era. This way to the egress folks come see the magnificent egress!

[–] AA5B@lemmy.world 1 points 3 days ago

Sometimes it seems like the biggest success of AI has been refining the definition of intelligence. But we still have a long way to go

[–] brie@programming.dev -5 points 4 days ago (4 children)

Large context window LLMs are able to do quite a bit more than filling the gaps and completion. They can edit multiple files.

Yet, they're unreliable, as they hallucinate all the time. Debugging LLM-generated code is a new skill, and it's up to you to decide to learn it or not. I see quite an even split among devs. I think it's worth it, though once it took me two hours to find a very obscure bug in LLM-generated code.

[–] cley_faye@lemmy.world 4 points 3 days ago (1 children)

If you consider debugging broken LLM-generated code to be a skill… sure, go for it. But, since generated code is able to use tons of unknown side effects and other seemingly (for humans) random stuff to achieve its goal, I'd rather take the other approach, where it takes a human half an hour to write the code that some LLM could generate in seconds, and not have to learn how to parse random mumbo jumbo from a machine, while getting a working result.

Writing code is far from being the longest part of the job; and you gingerly decided that making the tedious part even more tedious is a great idea to shorten the already short part of it…

[–] brie@programming.dev 0 points 3 days ago

It's similar to fixing code written by interns. Why hire interns at all, eh?

Is it faster to generate then debug or write everything? Needs to be properly tested. At the very least many devs have the perception of being faster, and perception sells.

It actually makes writing web apps less tedious. The longest part of a dev job is pretending to work actually, but that's no different from any office jerb.

[–] sudneo@lemm.ee 1 points 3 days ago (1 children)

Humans are notoriously worse at tasks that have to do with reviewing than they are at tasks that have to do with creating. Editing an article is more boring and painful than writing it. Understanding and debugging code is much harder than writing it etc., observing someone cooking to spot mistakes is more boring than cooking etc.

This also fights with the attention required to perform those tasks, which means a higher ratio of reviewing vs creating tasks leads to lower quality output because attention is depleted at some point and mistakes slip in. All this with the additional "bonus" to have to pay for the tool AND the human reviewing while also wasting tons of water and energy. I think it's wise to ask ourselves whether this makes sense at all.

[–] brie@programming.dev 0 points 3 days ago (1 children)

To make sense of that, figure out what pays more observing/editing or cooking/writing. Big shekels will make boring parts exciting

[–] sudneo@lemm.ee 1 points 3 days ago (1 children)

Think also the amount of people doing both. Also writers earn way more than editors, and stellar chefs earn way more than cooking critics.

If you think devs will be paid more to review GPT code, well, I would love to have your optimism.

[–] brie@programming.dev 1 points 3 days ago (1 children)

I'm too unfamiliar with the cooking and writing/publishing biz. I'd rather not use this analogy.

I can see many business guys paying for something like Devin, making a mess, then hiring someone to fix it. I can see companies not hiring junior devs, and requiring old devs to learn to generate and debug. Just like they required devs to be "full stack". You can easily prevent that if you have your own company. If ... Do you have your own company?

[–] sudneo@lemm.ee 1 points 3 days ago (1 children)

I don't, like 99% of people don't or won't. My job is safe, I am arguing from a collective perspective.

I simply don't think companies will act like that. Also the mere reduction of total number of positions will compress salaries.

[–] brie@programming.dev 1 points 3 days ago* (last edited 3 days ago) (1 children)

What collective perspective? There's gonna be winners and losers, non uniform rewards and costs. Companies are already acting like that. And IMO more will join. They're a hive mind who eagerly copy Google, Amazon, Facebook. And younger devs will add "LLM code gen" to their resumes. No job is safe, even kings and dictators get their heads chopped off.

[–] sudneo@lemm.ee 1 points 3 days ago (1 children)

Tech worker collective perspective. Those whose jobs will generally be affected due to idiots in position of power who are ready to believe that LLMs can do a good chunk of what devs do.

[–] brie@programming.dev 1 points 2 days ago (1 children)

What does it say about you if you let idiots have power over you?

[–] sudneo@lemm.ee 1 points 2 days ago (1 children)

That the world is not a cartoon and even idiots have structural power and it's not me "letting them"? Aldo again, this is not a "me" discussion. I will be fine. But many people in the industry will be screwed.

[–] brie@programming.dev 1 points 2 days ago (1 children)

It's similar to when factories got mechanized, and people were promised 2 day work weeks. In reality, the number of high paying jobs shrunk, and wages compressed. That's just the march of progress:)

[–] sudneo@lemm.ee 1 points 2 days ago (1 children)

Well, since I have the chance to at least fight against idiots who push against their own interests (my colleagues and peers) I do.

[–] brie@programming.dev 1 points 2 days ago (1 children)

So you'll be breaking Altman's beak with a baseball bat?

[–] sudneo@lemm.ee 1 points 2 days ago (1 children)

Do you think he is my peer or colleague? I wish...

Also Altman is not the problem, managers everywhere are

[–] brie@programming.dev 1 points 2 days ago

I just pictured Altman when I read "fight managers".

That's too many beaks for you to break. You'll have to wait 20 years for robotics to improve.

[–] spankmonkey@lemmy.world 2 points 4 days ago (1 children)

What is your favorite flavor of kool aid?

[–] NigelFrobisher@aussie.zone 1 points 3 days ago* (last edited 3 days ago) (1 children)

I have one of those at work now, but my experience with it is still quite limited. With Copilot it was quite useful for knocking up quick boutique solutions for particular problems (stitch together a load of PDFs sorted on a name heading), with the proviso that you might end up having to repair bleed between dependency versions and repair syntax. I couldn’t trust it with big refactors of existing systems.

[–] brie@programming.dev 1 points 3 days ago

Cursor and Claude are a lot better than Copilot, but none of them can be trusted. For existing large code repos, LLMs can generate tests and similar boring stuff. I suspect there'll be an even bigger shift to micro services to make it easier for LLMs generate something that works.