Google has introduced new real-time AI capabilities for its Gemini Live service, allowing it to see screens and camera feeds. This feature is currently available to select Google One AI Premium subscribers, marking a significant advancement in AI technology. The rollout follows nearly a year after the initial demonstration of Project Astra, which powers these capabilities.
The new features include screen-reading and live video interpretation, enabling users to ask Gemini questions about what it sees. For instance, users can seek advice on paint colors for pottery using their smartphone camera. This development highlights Google's competitive edge in AI, especially as rivals like Amazon and Apple are also enhancing their AI assistants.
• Gemini can now interpret screens and camera feeds in real-time.
• Google's AI advancements position it ahead of competitors like Amazon and Apple.
Real-time AI refers to systems that process data and provide responses instantly, as demonstrated by Gemini's new capabilities.
Screen-reading technology allows AI to interpret visual information from screens, enhancing user interaction.
Live video interpretation enables AI to analyze and respond to video feeds in real-time, facilitating dynamic user queries.
Google is a leader in AI development, with Gemini's new features showcasing its advancements in real-time processing.
Amazon is enhancing its AI capabilities with the upcoming Alexa Plus upgrade, competing directly with Google's innovations.
The Indian Express 11month
Isomorphic Labs, the AI drug discovery platform that was spun out of Google's DeepMind in 2021, has raised external capital for the first time. The $600
How to level up your teaching with AI. Discover how to use clones and GPTs in your classroom—personalized AI teaching is the future.
Trump's Third Term? AI already knows how this can be done. A study shows how OpenAI, Grok, DeepSeek & Google outline ways to dismantle U.S. democracy.
Sam Altman today revealed that OpenAI will release an open weight artificial intelligence model in the coming months. "We are excited to release a powerful new open-weight language model with reasoning in the coming months," Altman wrote on X.