this post was submitted on 03 Jun 2024
22 points (100.0% liked)
TechTakes
1401 readers
195 users here now
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
So apparently Mozilla has decided to jump on the bandwagon and add a roided Clippy to Firefox.
I’m conflicted about this. On the one hand, the way they present it, accessibility does seem to be one of the very few non-shitty uses of LLMs I can think of, plus it’s not cloud-based. On the other hand, it’s still throwing resources onto a problem that can and should be solved elsewhere.
At least they acknowledge the resource issue and claim that their small model is more environmentally friendly and carbon-efficient, but I can’t verify this and remain skeptical by default until someone can independently confirm it.
The accessibility community is pretty divided on AI hype in general and this feature is no exception. Making it easier to add alt is good. But even if the image recognition tech were good enough—and it’s not, yet—good alt is context dependent and must be human created.
Even if it’s just OCR folks are ambivalent. Many assistive techs have native OCR they’ll do automatically, and it’s better, usually. But not all, and many AT users don’t know how to access the text recognition them when they have it.
Personally I’d rather improve the ML functionality and UX on the assistive tech side, while improving the “create accessible content” user experiences on the authoring tool side. (Ie. Improve the braille display & screen reader ability to describe the image by putting the ML tech there, but also make it much easier for humans to craft good alt, or video captions, etc.)
I deleted a tweet yesterday about twitter finally allowing alt descriptions on images in 2022 - 25 years after they were added to the w3c spec (7 years before twitter existed) . But I added the point that OCR recommendations for screenshots of text has kinda always been possible, as long as they reliably detect that it's a screenshot of text. But thinking about the politics of that overwhelmed me, hence the delete.
Like, I'm kinda sure they already OCR all the images uploaded for meta info, but the context problem would always be there from an accessibility POV.
My perspective is that without any assistance to people unaware of accessibility issues with images beyond "would you like to add an alt description" leaves the politics of it all between the people using twitter. I don't really like seeing people being berated for not adding alt text to their image as if twitter is not the third-party that cultivated a community for 17 years without ALT descriptions, then they suddenly throw them out there and let us deal with it amongst ourselves.
Anyway... I will stick to what I know in future
read that back and it's a bit of an unreadable brain-dump. Apologies if it's nonsense
Yah, this makes sense. Community conventions can encourage good accessible content creation, and software can have affordances to do the same. Twitter, for many years, has been the opposite. Not only did it not allow alt, but the shorthand and memes and joke templates that grew up on short form Twitter was an extremely visual language. Emoji-based ascii art, gifs, animated gifs, gifs framed in emoji captioned by zalgo glitch unicode characters… there’s HTML that can make all that accessible, in theory, but the problem is more structural than that.
As a rough rule of thumb, if it's running locally on a CPU with acceptable performance, the environmental impact is going to be minimal, or at least within socially acceptable bounds.
This elides whatever resources were used in training/development. Which in the cases of ML models is quite often not minimal. Even the diy things you can train yourself make a significant dent. And there’s likely to be an ongoing cost of this too, because of updates