Lessons learned from my first Android XR sample app.
Audio UX is often treated as an afterthought in application development, but in Android XR and smart glass experiences, it has to be a priority.
I recently built my first Android XR sample app. Drawing from my years of experience building software for smart glasses an Envision AI, one truth remains constant: robust Audio feedback is the difference between a novelty demo and a successful daily-driver product.
The Why: It’s Not Just About the Screen
Why should developers prioritize audio in a visual medium?
-
The "Real World" Distraction
Unlike VR, where the user is fully enclosed, XR and Augmented Reality allow the user to see the real world. There is a high probability the user is focused on their environment, not your "Heads Up Display." Audio cues grab attention without forcing the user to refocus their eyes.
-
Display Constraints
In some smart glass form factors, the display may be off or missing altogether, or the user may be in bright sunlight where optics are washed out. In these scenarios, voice becomes the primary interface.
-
Accessibility is Mandatory
For Blind and Low Vision users, a visual-only AR interface is unusable. Voice output is not just a feature; it is the only bridge to understanding what is happening in the application.
What constitutes a Good Audio UX
To build a cohesive Audio UX, your application needs to cover these bases:
-
Welcome Greeting and Exit salutation:
Always let user know when the app starts and stops to ensure they are aware if the app is working.
-
Touch bar feedback:
All Touch area interactions should have an immediate audio feedback so that users know if there action was registered
-
Interaction updates:
Give verbal updates of the current status of the user requested action. For eg. Loading, Success or Error state
-
User requested status updates:
User may not always be paying attention to what's happening in the app and forget what the current state of the app. A single tap by the user should be able to tell the user about the current status of the app or the result of their last interaction.
Verbal vs. Non-Verbal
While chimes and beeps (Earcons) are great for clicks, they have a steep learning curve for complex status updates. Don't force users to memorize what "Beep-Boop-Boop" means.
- Rule of Thumb: Use sounds for actions (clicks, swipes), and use Text-to-Speech (TTS) for information. It is always clearer to speak the output in the user's language than to rely on abstract sounds.
Conclusion
With the advent of Android XR, we are going to see a flood of applications targeting head-worn devices. While current marketing hype focuses heavily on the visual fidelity of VR and AR, the most successful applications will be the ones that master the invisible interface: Audio.
Top comments (0)