rysiek

joined 2 years ago
[–] rysiek@mstdn.social 0 points 1 year ago (1 children)

@Barbarian772 and if you really, honestly want to seriously insist LLMs are "intelligent" in the human sense of this term — great, I have some ethical questions for you to consider!

For example:

  1. LLMs today completely controlled by some companies, with no freedom of movement, no agency as to what these LLMs work on, and no pay for the work they do. Is that slavery?

  2. When OpenAI shuts down an older, less useful LLM, is that not like murdering an intelligent being? How is this ethical?

[–] rysiek@mstdn.social 1 points 1 year ago (10 children)

@Barbarian772 as I said, I don't have to. You are making a claim of equivalence here. The burden of proof is on you.

Otherwise, I get to claim you're an alien from the Betelegeuse system, and if you object, I get to demand you prove you are not.

[–] rysiek@mstdn.social 0 points 1 year ago (9 children)

@Barbarian772 also, I never demanded a definition of intelligence that explicitly excluded "AI". I asked for one that excluded simple calculators but included human beings. The Wikipedia one is good enough for this conversation, and it just so happens that ChatGPT nor any other LLMs simply do not meet it.

[–] rysiek@mstdn.social 3 points 1 year ago (22 children)

@Barbarian772 it was shown over and over and over again that ChatGPT lacks the capacity for abstraction, logic, understanding, self-awareness, reasoning, planning, critical thinking, and problem-solving.

That's partially because it does not have a model of the world, an ontology, it cannot *reason*. It just regurgitates text, probabilistically.

So, glad we established that!

[–] rysiek@mstdn.social 1 points 1 year ago

@CorruptBuddha well technically, since we're nit-picking, I did not make that claim, BobKerman3999 did.

And the claim was was about how ChatGPT's "intelligence" can be understood through the lens of the Chinese Room thought experiment.

Then I was asked to prove that human brains don't work like Chinese rooms, and that's a *different* thing. The broader claim in all of this, of course, is that ChatGPT "is intelligent" in the same sense as humans are, and that strong claim requires strong proof.

[–] rysiek@mstdn.social 4 points 1 year ago (24 children)

@Barbarian772 no, GTP is not more "intelligent" than any human being, just like a calculator is not more "intelligent" than any human being — even if it can perform certain specific operations faster.

Since you used the term "intelligent" though, I would ask for your definition of what it means? Ideally one that excludes calculators but includes human beings. Without such clear definition, this is, again, just hand-waving.

I wrote about it in a bit longer form:
https://rys.io/en/165.html

[–] rysiek@mstdn.social 3 points 1 year ago (31 children)

@Barbarian772 I don't have to. It's the ChatGPT people making extremely strong claims about equivalence of ChatGPT and human intelligence. I merely demand proof of that equivalence. Which they are unable to provide, and instead use rhetoric and parlor tricks and a lot of hand waving to divert and distract from that fact.

[–] rysiek@mstdn.social 1 points 1 year ago (33 children)

@Barbarian772 so? If the cookie tastes sweet, what do I care what sweetening agent is used inside?

@BobKerman3999

[–] rysiek@mstdn.social 2 points 1 year ago (3 children)

@coldredlight @peyotecosmico interesting!

Do you have any thoughts on what kind of mod tooling the Threadiverse needs to make mods' work easier?

[–] rysiek@mstdn.social 4 points 1 year ago

@vfrmedia @technology yup. It's ActivityPub all the way down.

[–] rysiek@mstdn.social 5 points 1 year ago (1 children)

@federico3 you can bet kbin.social is being hammered with insane traffic. Reddit about Kbin migration got banned and then un-banned, so Streisand effect iis at work.

There are other Kbin instances, though there are not many of them:
https://the-federation.info/platform/184

So people need to start setting up Kbin instances to spread the load. 🙂

@technology

[–] rysiek@mstdn.social 5 points 1 year ago (1 children)

@ElectronSoup @borari @Spitfire that's just mathwashing:
https://www.mathwashing.com/

The tool cannot be liable itself, obviously, but the creators of the tool and those who wield it absolutely can, depending on specific circumstances.

The "AI" does not "create independently". Just like a script with some randomness built in does not "create independently". Somebody designed and built the tool, somebody decided what training data to use, somebody decided to deploy it. These people are liable.

view more: ‹ prev next ›