this post was submitted on 03 Dec 2023
461 points (99.1% liked)

Technology

59414 readers
3119 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Senate bill aims to stop Uncle Sam using facial recognition at airports / Legislation would eliminate TSA permission to use the tech, require database purge in 90 days::Legislation would eliminate TSA permission to use the tech, require database purge in 90 days

you are viewing a single comment's thread
view the rest of the comments
[–] NocturnalMorning@lemmy.world 38 points 11 months ago (2 children)

Facial recognition is bad for a multitude of privacy reasons. But, the biggest reason though is it is also wrong, and often trained with biased data (which is almost impossible to completely remove).

[–] bobgusford@lemmy.world 7 points 11 months ago (1 children)

Sorry, this needs more clarification! Do you mean "intent recognition" where some AI, trained with biased data, will assume that some brown person is upto no good? Or do you mean that they will misidentify black and brown people more often due to how cameras work? Because the latter has nothing to do with biased data.

[–] yeather@lemmy.ca 6 points 11 months ago (1 children)

Both in fact. Training data for things like this regularly mix up minority people. If Omar is a upstanding citizen, but gets his face mixed with Haani, known terrorist, Omar gets treated unfairly, potentially to the point of lethality.

[–] bobgusford@lemmy.world 3 points 11 months ago

For "intent recognition", I agree. A system trained on data of mostly black committing crimes might flag more black people with ill intent.

But for the sake of identification at security checkpoints, if a man named Omar - who has an eerie resemblance to Haani the terrorist - walks through the gates, then they probably need to do a more thorough check. If they confirm with secondary data that Omar is who he says he is, then the system needs to be retrained on more images of Omar. The bias was only that they didn't have enough images of Haani and Omar for the system to make a good enough distinction. With more training, it will probably be less biased and more accurate than a human.

[–] paysrenttobirds@sh.itjust.works -2 points 11 months ago (1 children)

There is nothing in the article to suggest that the TSA programs' errors have inconvenienced people as the agent is right there to correct it, and more scans improves the accuracy. I get what you're saying, but the same biases are undoubtedly programmed into the brains of the agents and just as hard to eradicate.

There are many places I don't want to see facial recognition employed, but where people are already mandated to positively identify themselves seems like a natural fit. I think the senators and the ACLU can find much more persuasive examples of overreach.

[–] NocturnalMorning@lemmy.world 4 points 11 months ago (1 children)

You're free to offer up being tracked and monitored everywhere you go, but I'm not okay with that.

[–] paysrenttobirds@sh.itjust.works 1 points 11 months ago

My point is the bill would be more interesting if it was not restricted to the TSA in an airport, but maybe they have to start somewhere?