this post was submitted on 20 Dec 2023
2 points (100.0% liked)

SneerClub

983 readers
16 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 1 year ago
MODERATORS
top 9 comments
sorted by: hot top controversial new old
[–] Soyweiser@awful.systems 1 points 11 months ago* (last edited 11 months ago) (1 children)

Basing your security on Movie plot threats is a pretty bad idea.

E: also somewhere in the NSA there is a junior analyst feverishly writing a report right now:

We should stop developing AI, we should collect and destroy the hardware and we should destroy the chip fab supply chain that allows humans to experiment with AI at the exaflop scale. Since that supply chain is only in two major countries (US and China), this isn't necessarily impossible to coordinate

(somebody on LW highlighted the last line as important (I also think the line is wrong but im not going to 'somebody is wrong on the internet' help the wannabe terrorist cultists)).

[–] skillissuer@discuss.tchncs.de 0 points 11 months ago (1 children)

this isn’t necessarily impossible to coordinate

Famously worked out with nukes

[–] Soyweiser@awful.systems 1 points 11 months ago* (last edited 11 months ago)

That is the one way to look at it.

The other way of looking at it is 'these are the only two places that build this, would be a shame if something happened to them'. Nukes could also work then, but I hear getting those is slightly tricky. (More of a sidestep governments go full Propaganda of the deed)

Of course Yuds T-shirt clearly says 'I do not want to nukedatacenters!(*)'

*: Conventional explosives also work.

[–] CodexArcanum@lemmy.world 1 points 10 months ago

We should also consider how the efforts of AI can be directed towards solving human aging; if aging is solved then everyone's time preference will go down a lot and we can take our time planning a path to a stable and safe human-primacy post-singularity world.

This shit is so funny. It always comes back around to immortality with these people.

I spent my adult life to this point seeking to understand enlightenment and transcendence, and if you aren't reading the doctrines of an immortality cult (lots of that in daoism and tescreal, oddly enough) then mostly you come to find that overcoming the fear of death is a big part of it. You can't have a free and open mind if the shadow of your mortality looms, so you learn to let go of it as one more attachment.

I think it's very revealing of what shallow minds these people truly have that the desperate craving for immortality is so naked in their beliefs. That the concept of it goes so unexamined as well (oh we just make everyone immortal and then we all agree to debate and solve our AI problem? Because that would work?)

Like, how stupid do you have to be to say on the one hand that unexamined AI risk requires a massive effort to run simulations and game theory out the consequences, but then on the other hand to be like "things would be better if everyone was immortal. I will take no further questions on that."

[–] 200fifty@awful.systems 1 points 11 months ago* (last edited 11 months ago) (2 children)

we simply don't know how the world will look if there are a trillion or a quadrillion superhumanly smart AIs demanding rights

I feel like this scenario depends on a lot of assumptions about the processing speed and energy/resource usage of AIs. A trillion is a big number. Notably there's currently only about 0.8% this number of humans, who are much more energy efficient than AIs.

[–] bitofhope@awful.systems 1 points 11 months ago (1 children)

During WWII everyone computed on slide rules which had zero transistors. Then they invented the transistor which had one transistor. Then they started making mainframes and Ataris and C64s which had like what, hundred or thousand? Then they invented computers and Windows and PS1 that had maybe a million transistors. And then we got dual core CPUs which had double the transistors per transistor. Then they invented GPUs which is like a thousand tiny CPUs in one CPU. Then they made i7 which probably has like a billion transistors and Ryzen which has ten billion and RTX4090 Ti has 79 billion. Now they say China is going to make a phone with trillion transistors.

That's called exponential growth and assuming perfectly spherical frictionless nanometers it will go on forever. In eight to twelve years schoolchildren will be running GPT6 on their calculators. We will build our houses entirely out of graphics cards. AI will figure out cold fusion any week now and Dennard scaling will never hit its limit, right?

[–] locallynonlinear@awful.systems 1 points 11 months ago (1 children)

A trillion transistors on our phones? Can't wait to feel the improved call quality and reliability of my video conferencing!

[–] Evinceo@awful.systems 1 points 11 months ago

now you're in pure fantasy land.

[–] locallynonlinear@awful.systems 1 points 11 months ago

We simply don't know how the world will look X (anything with a bigger scale)

Yes. So? This has, will, always be the case. Uncertainty is the only certainty.

When these assholes say things, the implication is always that the future world looks like everything you care about being fucked, you existing in an imprisoned state of stasis, so you better give us control here and now.