Google details how users will control and navigate Android XR glasses ahead of 2026 launch

HIGHLIGHTS

Android XR will come in 2 form factors: AI Glasses and Display AI Glasses.

Android XR glasses will support touch, voice and hardware button controls.

New 'Glimmer' design language guidelines are set for optical see-through displays.

Google details how users will control and navigate Android XR glasses ahead of 2026 launch

Google has been releasing detailed design documentation and development tools related to its forthcoming Android XR glasses platform. 9to5Google has found updated instructions that tell how users will be able to control and navigate these glasses. Read on to know what to expect in terms of their hardware controls, software navigation, display behaviour and thermal dissipation. This should give you an idea of what the overall Android XR experience will look like.

Digit.in Survey
✅ Thank you for completing the survey!

Two form factors: AI Glasses and Display AI Glasses

Google has officially named the two hardware categories:

  • AI Glasses will include speakers, a microphone and a camera, but no built-in display.
  • Display AI Glasses will add a small screen. Single-display models are referred to as monocular, while dual-display versions will be referred to as binocular models. The binocular variants are expected later.

Importantly, users can turn off the display at any time. This means you can have visuals when needed or enjoy voice-only when preferred. It also indicates that Google sees voice and AI interaction, likely powered by Gemini, as central to the platform rather than just an add-on.

Also Read: Best smart glasses with AI, AR and VR you can buy in India

Hardware controls: buttons, touchpad and LEDs

Google is standardising key physical controls across all Android XR glasses. Every device must include:

  • A power switch or button
  • A touchpad
  • A dedicated camera button

Models with displays will also feature a separate display button, typically placed on the underside of the stem. It allows users to wake or sleep the screen, effectively switching between visual and audio-only modes.

The camera button supports multiple actions. A single tap captures a photo, while a long press records video. Pressing again stops recording, and a double press launches the Camera app.

You can use the touchpad through these gestures:

  • Tap to play, pause or confirm
  • Touch and hold to invoke Gemini
  • Swipe for navigation
  • Two-finger swipe to adjust volume
  • Swipe down to go back or return to the home screen

On display models, swipes also handle scrolling, focus movement and button selection inside apps.

Each device will include two LEDs, one visible to the wearer and one to bystanders. These act as system indicators for whether the device is on, off, whether it is recording, etc. Developers cannot modify these LEDs. So, you see, Google is standardising privacy cues across the ecosystem.

Software: home screen, system bar and multitasking

On Display AI Glasses, users will see a home screen that Google compares to a smartphone lockscreen. At the bottom sits a persistent system bar showing time, weather, notifications, alerts and Gemini feedback.

Above it, users can view contextual, glanceable information, suggested shortcuts to likely next actions, and ongoing activities when multitasking.

Notifications appear as pill-shaped chips, and you can select to expand them.

Google has chosen ‘Glimmer’ design language for Android XR glasses with guidelines for soft corners, what colours to choose based on power and heat efficiency, Material Symbols Rounded as the standard icon, and the use of a stack layout that shows one piece of content at a time.

So, you can expect consistency across brands employing the Android XR platform. Google’s approach includes a combination of physical buttons, touch gestures and voice control in a lightweight glasses format. By requiring apps to function without a display, Google is prioritising AI-driven, voice-first interaction as a core experience.

Let’s see how the end products pan out. Keep reading Digit.in for similar updates.

Also Read: Meta smart glasses may get facial recognition feature: Report

G. S. Vasan

G. S. Vasan

G.S. Vasan is the chief copy editor at Digit, where he leads coverage of TVs and audio. His work spans reviews, news, features, and maintaining key content pages. Before joining Digit, he worked with publications like Smartprix and 91mobiles, bringing over six years of experience in tech journalism. His articles reflect both his expertise and passion for technology. View Full Profile

Digit.in
Logo
Digit.in
Logo