[BRLTTY] Is there a feature-compatible text-based browser
Kyle
kyle at gmx.it
Mon Sep 29 11:12:13 UTC 2025
I do find all this AI stuff to be very interesting indeed, although as
has been said, I can't see it replacing the screen reader entirely; it
may work better as a supplemental tool, much like the way BeMyAI
complements and supplements the BeMyEyes volunteer. As it stands now, I
can actually ask several open source AI models to describe a picture
taken with my phone's camera and get a halfway decent response back in
about the same time as it takes to upload the same picture to BeMyEyes
and wait for its AI to come back with a response.
The main problem with the AI replacing the screen reader though is not
speed, but hallucination, which is still a huge problem with every model
I've ever used for any purpose. I mean BeMyAI described one of the 2000
gold dollar coins as a giant penny, complete with Lincoln's face and
all. But I knew it was hallucinating, because I knew exactly what coin I
was holding. Many times when the AI hallucinates, we don't know that is
what is happening.
And the bigger problem is that I don't want my computer to try to think
for me. My workflow is pretty straightforward. I want to do something, I
either look in the menu or type in a command to find it, the computer
does it. Or I'm on a website, I want to see what is on the page, and if
I'm lucky, the headers are marked so that I at least get a nice idea.
And if I want a page summary before I get started, my screen reader can
do that at the press of a button. I don't see any benefit of AI here,
with the obvious exception of text recognition or helping to map out an
otherwise inaccessible window so that its characteristics can be sent to
the screen reader, which could then read what was sent to it by the AI
as I interact with it normally. I don't want a detailed description of
the whole window, only the control I'm focusing on at the time and any
text that may need my attention when the window pops up. AI descriptions
now are still a bit too wordy, sometimes leading to additional confusion
rather than a straightforward workflow.
Yes, I for one enjoy the graphical desktop and the consistency that it
provides, i.e. one key combination has its function everywhere instead
of all these little programs having different key sequences that all end
up doing the same thing; e.g. if I press a q here, it closes the
application, but if I press it in another application, nothing happens,
and I was supposed to use control+x, which incidentally is the cut
command *everywhere* on the MATE desktop and GNOME as well. I notice
nothing slow about writing this message in Thunderbird, which used to be
pretty darn slow just 10 or so years ago, but is smooth now, and this is
on a laptop that is about 8 years old and a desktop that is about 12. So
I can see how in the future, AI may become useful enough locally that it
can be just as fast as interacting with a graphical desktop is now. But
then most AI models still rely too heavily on the GPU, so this may come
at some point down the road, not I fear in the next year or so, but
maybe in the next 10. Still, the above problems of hallucination and its
attempts to think for me are still a bit off-putting to me, even if they
can fix the lag problems it would introduce now.
~Kyle
More information about the BRLTTY
mailing list