Nielsen/Norman Group’s Katie Sherwin just posted an article
discussing the experience of a blind person trying to interact with a
touchscreen device and its rich vocabulary of gestures. The complexity of the
interface, plus the many gestures available can create quite a “cognitive load”
as the blind user tries to maintain an idea of what is happening on the screen
to know what gestures to use.
While some products have experimented with tactile/haptic
interfaces that give physical feedback, the potential of such designs is limited.
The better answer, Sherwin argues, to simplify content and workflow so people
can accomplish their goals more efficiently.
Screen Readers on Touchscreen devices