Building on our Gemini models, Project Astra explores the future of AI assistants that can process multimodal information, understand the context youâ

Project Astra - Google DeepMind

submited by
Style Pass
2024-05-16 00:30:05

Building on our Gemini models, Project Astra explores the future of AI assistants that can process multimodal information, understand the context you’re in, and respond naturally in conversation.

Building on our Gemini models, Project Astra explores the future of AI assistants that can process multimodal information, understand the context you’re in, and respond naturally in conversation.

This demo shows two continuous takes: one with the prototype running on a Google Pixel phone and another on a prototype glasses device.

To be truly useful, an agent needs to understand and respond to the complex and dynamic world just like people do — and take in and remember what it sees and hears to understand context and take action. It also needs to be proactive, teachable and personal, so users can talk to it naturally and without lag or delay.

While we’ve made incredible progress developing AI systems that can understand multimodal information, getting response time down to something conversational is a difficult engineering challenge.

Leave a Comment