Google has begun rolling out new real-time AI video features for Gemini, its AI assistant, allowing it to interpret and respond to content on users' smartphone screens and camera feeds. This development is part of Google's Project Astra, which was first demonstrated nearly a year ago. The features are currently available to Gemini Advanced Subscribers of the Google One AI Premium plan, with initial rollout targeting Pixel devices, some Samsung phones, and other Android devices including Xiaomi phones. The new capabilities enable Gemini to provide immediate, context-aware assistance based on visual input. Users can point their phone cameras at objects or scenes, and Gemini will analyze the visual data to offer relevant information or answer questions in real-time. This marks an advancement in integrating AI into everyday tasks, enhancing user interaction with technology. In addition to screen and camera analysis, Google is preparing to unveil a new model for Gemini this week, which will include a toolbox for 'Agents.' These agentic use cases are expected to further enhance Gemini's capabilities, potentially integrating with existing features like Canvas and Deep Research.
GPT-4o Image Generation Today 11am PT https://t.co/RSa5aSylxJ
Big AI news day today: New @Google Gemini 2.5 model is rolling out that integrates their Imagen 3 model into main chat @OpenAI responding with an new image model live stream at 11am PST https://t.co/dqNMV63qda
Google DeepMind's Gemini Robotics and Gemini Robotics-ER AI models are revolutionizing robotics with advanced vision, language, and spatial reasoning capabilities. #datascience #AI #artificialintelligence https://t.co/MAHbr9tUe5