Before an important interview or on a nerve-wracking first date, does your heart ever feel like it’s pumping at a million beats per minute? Stanford University students have come up with a solution to put this common problem to rest – making sure that no one has to worry about what to say.
Stanford student Bryan Hau-Ping Chiang has recently brought new technology to the forefront with a prototype of glasses that “listen” to conversations and suggest the next words to say. Chiang tweeted about this revolutionary invention.
In a video demonstration of the AI/AR glasses, Varun can be seen wearing the monocle and being posed questions by Alix. Stanford researchers were delighted by the capabilities displayed in this demonstration.
With a slight delay, the glasses interpret the question and transcribe the answer onto their screen. Varun then reads out that answer so that it can be heard.
The Monocle AR glasses, provided by Brilliant Labs and equipped with OpenAI’s Whisper, a large language model for speech recognition, and a microphone, camera, and high-resolution display, are showcased in the given snapshot from the demonstration.
Understand How Your Glasses Work – A Guide For Beginners
Using Bluetooth, ‘rizzGPT’ AI glasses seamlessly connect with a web application already on the host device. This allows audio to be changed in real-time into text format for translation when dialogue is initiated.
OpenAI’s Whisper technology, employed with its glasses, allows for conversation to be processed by a chatbot. After being triggered by queries from the chats, the chatbot will respond with pertinent information, which is then relayed back to the user wearing them.
The development of AI-powered glasses prompts important discussions about the role of technology in our lives and the potential impact on social dynamics. As this technology continues to evolve, it will be crucial to consider the ethical implications and strive for a balance between the benefits and potential drawbacks.