Shahram Izadi, Google’s lead on the project, recently demoed the device at TED. While the prototype still lacks a catchy product name, its technical ambitions are clear: a conventional-looking frame that discreetly houses a microdisplay, camera, battery, speakers, and a robust AI assistant. No visible cables, no bulky modules—just a sleek form factor that offloads heavy computation to a paired smartphone.

Empfohlener redaktioneller Inhalt
Dieser externe Inhalt von YouTube wurde von der nextpit Redaktion ausgewählt und ergänzt den Artikel. Du hast die Wahl, ob du diesen Inhalt laden möchtest.
Externen Inhalt erlauben
Ich bin damit einverstanden, dass externer Inhalt geladen wird. Personenbezogene Daten werden womöglich an Drittplattformen übermittelt. Nähere Informationen enthält die Datenschutzerklärung.

Display, Optics & System Architecture

Under the hood, Google’s AR glasses feature a microdisplay embedded in the right lens, while the left front arm discreetly houses a camera. All compute-intensive tasks—such as AI inference and AR rendering—are offloaded to the connected smartphone. This architecture keeps the glasses lightweight and energy-efficient.

Gemini Memories: Contextual AI That Actually Remembers

The real game-changer? Google’s Gemini AI, now enhanced with “Memories”—a persistent, context-aware layer that logs and recalls both visual and auditory information. Unlike traditional assistants, Gemini doesn’t just respond to questions; it proactively observes, indexes, and retrieves data from your environment.

In the TED demo, Gemini identified book titles on a shelf and the location of a hotel keycard—without being explicitly prompted. This persistent memory layer marks a major leap in contextual computing, blurring the line between passive observation and active assistance. Imagine asking, “Where did I last see my passport?” and receiving a precise, visual answer. That’s the vision—and it’s closer than you might think.

Multilingual Mastery: Real-Time Speech and Text Translation

Google is also pushing boundaries with Gemini’s language capabilities. The glasses can recognize and translate text in real time—no surprise there—but also handle live speech in multiple languages, including Hindi and Farsi. This isn’t just basic translation; it’s full-duplex, context-aware communication, powered by on-device AI models.

Display Limitations: Where the Prototype Still Falls Short

Not everything is market-ready just yet. The current microdisplay, while functional, suffers from a limited field of view, edge distortion, and low brightness. During the demo, navigation cues and 3D city maps appeared small and pixelated—especially around the periphery.
Google’s roadmap clearly points to iterative improvements in optics and rendering, but don’t expect miracles—at least not in the first commercial release.