Okay, are they actually insane?
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
yes
For the record, I'm not super worried about AI taking over because there's very little an AI can do to affect the real world.
Giving them guns and telling them to shoot whoever they want changes things a bit.
So, it starts...
This is the best summary I could come up with:
The deployment of AI-controlled drones that can make autonomous decisions about whether to kill human targets is moving closer to reality, The New York Times reported.
Lethal autonomous weapons, that can select targets using AI, are being developed by countries including the US, China, and Israel.
The use of the so-called "killer robots" would mark a disturbing development, say critics, handing life and death battlefield decisions to machines with no human input.
"This is really one of the most significant inflection points for humanity," Alexander Kmentt, Austria's chief negotiator on the issue, told The Times.
Frank Kendall, the Air Force secretary, told The Times that AI drones will need to have the capability to make lethal decisions while under human supervision.
The New Scientist reported in October that AI-controlled drones have already been deployed on the battlefield by Ukraine in its fight against the Russian invasion, though it's unclear if any have taken action resulting in human casualties.
The original article contains 376 words, the summary contains 158 words. Saved 58%. I'm a bot and I'm open source!
I hope they put some failsafe so that it cannot take action if the estimated casualties puts humans below a minimum viable population.
Of course they will, and the threshold is going to be 2 or something like that, it was enough last time, or so I heard
The only fair approach would be to start with the police instead of the army.
Why test this on everybody else except your own? On top of that, AI might even do a better job than the US police
Cool, needed a reason to stay inside my bunker I'm about to build.
I'm guessing their argument is that if they don't do it first, China will. And they're probably right, unfortunately. I don't see a way around a future with AI weapons platforms if technology continues to progress.
It will be fine. We can just make drones that can autonomously kill other drones. There is no obvious way to counter that.
Cries in Screamers.
If we don’t, they will. And we can only learn by seeing it fail. To me, the answer is obvious. Stop making killing machines. 🤷♂️
Netflix has a documentary about it, it's quite good. I watched it yesterday, but forgot its name.