These notes augment my 2-hour ‘Lived Experience’ workshop around accessibility and voice experience design.
My UX viewpoints come from my three key perspectives/interest areas which are:
I’ll be making additions to these notes towards a conference presentation for later in the year.
Below are some ideas which we touched upon lightly or which we didn’t have time to explore in the session. Please let me know if you would like any clarifications or expansions.
I’ll explain my work-flow in audio only.
There are ideas in this that may help in voice app design, and understanding accessibility issues better.
My ears and brain get fatigued, we need to find silence amongst the noise of the world.
Accessibility = usability. Accessible Inclusive Design – AID
How do you decide what to put in a standard – it assumes we know what works best.
Distilling experience and understanding into guidelines is a key role in my past work.
Long Form Audio and Information
iVote by Phone: fully Automated Telephone Voting system in the NSW State Election
Today’s News Now: An automated service for browsing and reading (hearing) newspapers over the phone.
Both are examples of Inclusive Service Design
I’ll bring my portable office and run through how I manage various tasks
I’ll adorn myself with the wearables and peripherals that I use during the day and how they interact with one-another.
I’ll speak about the Apple ecosystem and what about it works for me and why
I’ll discuss the impact of touch-screens on blind people and demonstrate my RIVO – 2.0 an alternative input & output device for iOS devices.
AirPods (two sets) and Apple Watch.
Also Victor Reader Stream – talking book player
Braille Sense Mini – Braille input output note-taking device.
These notes and webpage contents were prepared and edited entirely in Markdown.
Bone Conduction headphones – Blue’s Titanium
Noise-cancelling and hear-through – Sennheiser Ambia Pro
Augmented Reality and staying connected with the outside world.
Near-field and far-field voice communication.
Occlusion shutting world out and making your voice boomy and distracting.
I’ll speak briefly about the current state-of-play for mobile apps iOS and Android, Web, ARIA and pitfalls of tools and pre-made Javascript libraries.
Google’s Flutter Deve kit mentions accessibility across iOS and Android platforms.
Will also touch on self-voicing applications vs apps that work with a screen-reader.
Apple TV Youtube and Amazon apps are examples of breaking Apple UI guidelines, leading to accessibility complications.
Voice First is the Amazon catch-cry but all but low-hanging fruit is shunted to a screen-based app. – Accessibility and usability of the Amazon Alexa app or the Google Home app for installation or configuration is arduous and in no way mirrors the voice simplicity of the device itself.
Misguided personality programming doesn’t properly prioritise the most relevant information.
Now that Siri is launched on the HomePod, which has no screen, Apple’s multi-modal touch-first model starts to show its biases.
Siri on HomePod trying to decide which device to respond Poor implementation. HomePod wins, even when it can’t execute the task.
Accessibility side-effect is that iPhone loses its place if in a list of items or mid-way through a long document read.
Device should be addressable e.g. HomePod, play some cool music.
Apple is presently leading the way in natural-sounding speech output, at a high audio quality. Over time this investment will strengthen the connection between a user and their assistant.
Google is working fast to create very human-sounding speech too.
Hearing Robotic stilted speech leads to a tendency to generate stilted speech in response – the foreigner in a strange country syndrome.
We are likely to think that If a system speaks funny, then probably it doesn’t ‘get me’. .
Courteous engagements. Do you say Please to your assistants?
Alexa and Google are looking at “the magic word” but in book Flutter it suggests that your speaking style and courteousness tells machines too much about us and our weaknesses.
Note that depending on the platform, these are usually issues beyond the control of a voice application designer, but longer-term these biases and factors need to be considered and addressed.
Is your choice of output voice aligned with your user base? Gender, ethnicity, accent, age, and personality? Do they feel included or separate?
Is your app’s purpose and audience suited to its voice? As a hypothetical, Is a female voice (on all the leading assistants) going to work for a voice-oriented gay male dating app akin to Grindr?
Is your speech recognition engine able to handle speech impediments, stutters, nervous speech, shaky and broken speech? The Mozilla Speech Recognition Corpus Project may be an opportunity to include folks with different speech profiles, speech impediments etc.
Are your timeouts sufficient for people who speak slowly or take more time to formulate their requests/responses?
Does your service have understanding of and respond to colloquial informal terms and phrases from your users? This also has a bearing on how comfortable and accepted your users feel.
Though not immediately apparent to everyone, eye-hand coordination should not be a design requirement for voice assistants, as they principally work in the auditory (non visual) domain.
Google Home Mini and HomePod both employ touch-sensitive controls for volume adjustment, pause/play etc.
Alexa in contrast has physical buttons – with nominal tactually differentiated surfaces, so it can be operated by feel, in the dark. – Though better design overall, physical buttons could be more problematic for people with physical disabilities to operate.
We are somewhere around v0.1 when it comes to voice assistants. This is why none of them reveal their version number. They are a service, not user software.
Blind from birth, Tim Noonan is a voice experience designer, inclusive design consultant and an expert in voice & spoken communication.
Building on his formal background in cognitive psychology, linguistics and education, Tim has been designing and crafting advanced voice interfaces since the early 90s and was one of the principle authors of the Australian and New Zealand standard on interactive voice response systems AS/NZS 4263.
Tim is the principle author of several other standards relating to automated voice systems, including automated telephone-based voting, telephone-based speech recognition and four industry standards on the accessibility of electronic banking channels and inclusive authentication solutions.
Tim has also been a pioneer in the accessibility field for more than three decades. He particularly loves working with emerging and future technologies to come up with innovative ways to make them more inclusive, effective and comfortable to use.
A career highlight for Tim was working as the lead Inclusive User Experience designer for iVote – a fully automated telephone-based and web-based voting system for the NSW Electoral Commission. iVote was issued with Vision Australia’s Making A Difference Award and was recommended as the ‘Gold Standard’ for accessible voting.
For the last 25 years Tim has been leading the way in teaching, conceptualizing and designing technologies that communicate with users through voice and sound – both for accessibility and mainstream users.
website by twpg