Comment by shakna

Comment by shakna 16 hours ago

11 replies

Whispir is a much better TTS than almost anything else. However, when it gets it wrong, oh boy does it get it wrong.

For everything else? Not really. JS thrashing the DOM is as much a pain as ever. Using ico files instead of either emoji or... Text... Makes UIs painful and inconsistent.

Everyone using Electron and its broken [0] accessibility, including core Windows features...

These aren't things that can be reasoned away with an LLM. An interface is not just text - its a reasoned nodegraph. And when I'm blind (comes and goes), I need the nodegraph. Not an image of the screen reinterpreted.

[0] https://github.com/electron/electron/issues/45856

graemep 16 hours ago

I find it very hard to know what to do to follow best practice. For example the biggest UK charity for blind people make social media posts about the importance of text descriptions and alt tags that break what I thought was good practice (they duplicate text in post and alt tag) and they seem to encourage this.

  • shakna 15 hours ago

    I don't recall where, but I've heard that before in the past. Perhaps in the kind of slop that makes the rounds on LinkedIn.

    There is sort of a good reason for it, in the past. Before the overhaul, Microsoft Speech used to skip Facebook posts, and read the alt text instead. It is now, however, more sane. Facebook was pretty darn bad at accessibility in its early days. A lot of intermingled broken spans for text, causing stuttering and other issues.

    However, today, most reading systems prefer the "title" attribute, to the "alt" one. If title exists, it'll read that and skip alt. Some always skip alt, regardless of it exists or not.

    Figure and figcaption are about the only way to get good and consistent behaviour, but you don't really control how those happen on most social media platforms. You throw everything you can at the wall, and see what happens. And it might change tomorrow.

    Today, I'd say the above is bad advice. An image description is a good practice. Repeating yourself, isn't.

    • graemep 13 hours ago

      Thank you, that is quite informative.

      The specific posts I see are from the Royal National Institute of Blind People who really ought to know.

      What they do is add the image description at the end of each text post, even thought this matches the alt text.

      This is the one about using alt text: https://www.facebook.com/rnibuk/posts/pfbid037RmtoSxfAJX82G4...

      They do now have a comment on that one that explains their reasoning (I did not see it until just now).

    • zersiax 14 hours ago

      Wait the title attribute?

      What are you basing that on? Screen readers tend to not pick those up at least on interactive elements by default, you need to do a bit of "wiggling" to get those to be announced. Disclaimer: screen reader user

      • shakna 14 hours ago

        JAWS user, here. It will read both aria-label and title, on a button, which is an interactive element. [0]

        It does depend on the verbosity, if you dial that down, you'll probably lose the title element. But for images, which is what I was mentioning, it should pretty much always be read out.

        [0] https://github.com/FreedomScientific/standards-support/issue...

jmward01 15 hours ago

I imagine this is where LLMs could really help actually. LLMs are natively surfing the web now so I suspect LLM descriptions of sites or even having them re-render a site in a more usable way is becoming much more possible.

  • UqWBcuFx6NV4r 15 hours ago

    Visually impaired people are more than willing to tell everyone what they need as far as accessible UIs, myself included. Barely anyone listens. Let’s not let LLMs be the next thing that people that don’t understand the actual problem try shoving in as a solution.

    Wheelchair users / people with mobility impairments rightfully scoff at the myriad “concepts” you see now and again of mech suits, wheelchairs that can climb stairs, etc. “Just give us a ramp! This is a solved problem. Your alternative is just sci-fi fetishism!” Still, it keeps happening. LLMs are increasingly becoming the same thing for people with visual impairments.

  • shakna 15 hours ago

    I don't need the text of the page. Thats easy, and I already have it.

    But information has a hierarchy, usually visual, and that hierarchy needs to be reflected. LLMs are famously bad at structure, especially any tree with significant depth. RAG is not enough - hallucinations become common at depth.

    My response now, to you, is in a semi-structured node graph. I know a reply has happened, because of the dangling children. I know who made it, and what they said, by cell attributes in the spans, surrounding it.

    Don't worry - AI is being shoved down accessibility's throat, like everywhere else. FSCompanion for JAWS, NVDA has an OpenAI plugin, and VoiceOver has it builtin.

    Why do I hate it? Because when it _doesn't work_, you can't tell. You don't know if it is hallucinating data, and cannot verify the response. If it is the mode of communication, it is all you have, making every failure a catastrophic failure.

    • jmward01 15 hours ago

      Thanks for helping me, and hopefully others, understand the challenges more!