Forget Google Glass. Google has instant translation specs
(Image credit: Future)
Yes, you heard that right. During Google I/O 2022, Google finally blew our minds with eyeglasses that literally translate what someone is saying to you before your eyes — provided you’re wearing the glasses, of course.
The glasses look perfectly normal, appear to work without any assistance from a phone, and almost have us forgetting all about Google Glass.
In truth, not much is known about the glasses beyond the short video demonstration, which Google ran at the tail end of a nearly two-hour event — a Steve Jobsian, “one more thing moment.”
The glasses use augmented reality and artificial intelligence (and possibly embedded cameras and microphones) to see someone speaking to you, hear what they’re saying, translate it and display the translation live on the embedded, translucent screens built into the eyeglass frames.
“Language is so fundamental to communicating with each other,” explained Google CEO Sundar Pichai, but he noted that trying to follow someone who is speaking another language “can be a real challenge.”
The Google Translation Glasses POV as depicted by Google (Image credit: Future)
The prototype lenses (thankfully, Google did not call them “Google Glass 3”) uses Google’s advancements in translation and transcription to deliver translated words “in your line of sight,” explained Pichai.
In the video, a young woman explains how her mother speaks Mandarin and she speaks English. Her mother understands her but can’t respond to her in English.
The young woman dons the black, horn-rim-style glasses and immediately sees what the researcher is saying, transcribed in yellow English letters on a screen. Granted, we’re seeing an overlay of this in the video and not what the woman is actually seeing.
What is clear, though is that no one is, as they had to with Google Glass, glancing upward to see an oddly-placed, tiny screen. The prototype lets wearers look directly at the speaker, so the words are overlayed on top of them. In one segment, we see a point-of-view image of the translation at work. Again, this is Google’s illustration of the prototype view. We won’t know how they really work until the prototype leaves the lab.
Google offered no time frame for completing and shipping this project. We don’t even have a name. Still, seeing normal-looking augmented reality glasses that could solve a very real-world problem, (they could translate sign language for someone who doesn’t know it or show words to the hearing impaired) is exciting.
As one researcher noted in the video, “Kind of like subtitles for the world.”