Demis Hassabis, head of Google DeepMind, foresees significant job disruption due to AI within 5-10 years, urging proactive skill development and adaptation, especially for the younger generation.
Google unveils Gemini 2.5, a significant AI model upgrade enhancing learning with LearnLM, offering advanced enterprise solutions like 'Deep Think' on Vertex AI, and providing state-of-the-art video understanding capabilities across its products.
Kawasaki unveils Corleo, a rideable robot horse concept powered by hydrogen. This AI-guided quadruped aims for future off-road mobility, targeting 2050, but its depicted abilities are currently speculative.
Gemini 2.5 Unveiled: Google's Next-Gen AI Enhances Learning, Enterprise, and Video Understanding
Google unveils Gemini 2.5, a significant AI model upgrade enhancing learning with LearnLM, offering advanced enterprise solutions like 'Deep Think' on Vertex AI, and providing state-of-the-art video understanding capabilities across its products.
Google integrates specialized LearnLM models directly into Gemini 2.5, significantly enhancing the AI's capabilities in personalized learning and explaining complex topics.
Gemini 2.5 Pro and Flash deliver state-of-the-art video understanding, enabling the transformation of videos into interactive applications, the generation of animations, and the precise analysis of video content.
New features in Google products like NotebookLM (Audio and Video Overviews), Search (AI Mode, Search Live), and the Gemini app (quiz creation) expand user capabilities.
Enterprise versions of Gemini 2.5, available via Vertex AI, offer new tools such as "Thought summaries" for improved auditability and "Deep Think mode" for solving highly complex tasks.
Developers and users can try out the new capabilities of Gemini 2.5 through Google AI Studio, the Gemini API, and the Vertex AI platform.
The landscape of artificial intelligence is continually shifting, with new developments rapidly integrating into the tools we use for learning, work, and daily interaction. Google has recently announced significant advancements to its AI model, Gemini 2.5, aiming to make knowledge more accessible and AI tools more powerful. These updates span enhanced learning capabilities through LearnLM, sophisticated enterprise solutions on Vertex AI, and substantial progress in video understanding, affecting a wide array of Google products and services.
LearnLM: Elevating the Learning Experience with Gemini 2.5
Google's commitment to making learning more active and effective is underscored by the infusion of LearnLM, its family of models fine-tuned for education, directly into Gemini 2.5. Ben Gomes, Chief Technologist, Learning & Sustainability at Google, stated, "Making knowledge accessible to everyone has always been our highest priority... With AI, we can do this at a speed and scale never before possible."
According to Google, Gemini 2.5 Pro, powered by LearnLM, outperformed competitors on every category of learning science principles in their latest report. This means Gemini can go beyond merely providing answers, helping to explain the process and untangle complex topics. A new prompting guide is available to demonstrate these capabilities.
These capabilities are being integrated across Google's product suite:
Google Search: The AI Mode in Search, rolling out to everyone in the U.S., features advanced reasoning and multimodality. Soon, Deep Search will allow for deeper exploration of complex topics. A custom version of Gemini 2.5 will power AI Mode and AI Overviews in the U.S. Furthermore, Search Live, integrating Project Astra's live capabilities, will allow users to ask questions about what they see in real-time. Search Live is coming to Labs this summer.
0:00
/0:24
Gemini App for Students: Free Gemini upgrades (Google AI Pro plan for 15 months, 2TB storage, NotebookLM) are now extended to college students (18+) in Brazil, Indonesia, Japan, and the United Kingdom who sign up by June 30, 2025. Globally, students can now create custom quizzes on any topic or uploaded documents.
0:00
/0:11
Experimental Learning Tools:
Project Astra: This research project is prototyping a conversational tutor for homework help, capable of step-by-step guidance and generating diagrams. Android Trusted Testers can sign up for the waitlist.
Learn About: This Labs project (sign up) now delivers more nuanced explanations, session history, and the ability to upload source documents for grounded explanations.
Sparkify: A new Labs experiment (waitlist available) that helps turn questions or ideas into short animated videos using Gemini and Veo models.
NotebookLM: This research and learning tool now features Audio Overviews in over 80 languages with selectable lengths for summaries, and Mind Maps. Coming soon are Video Overviews, which will turn notebook content into educational videos.
Gemini 2.5 for Enterprise: Driving Sophistication and Security
Google is also extending the power of Gemini 2.5 to enterprise users. Gemini 2.5 Flash will be generally available on Vertex AI and Google AI Studio in early June, with Gemini 2.5 Pro following soon after. These models introduce several key features for building sophisticated and secure AI-driven applications:
Thought Summaries: This feature organizes a model’s raw thoughts into a clear format, aiding in validation, alignment with business logic, and debugging for complex AI tasks.
Deep Think Mode: Utilizing new research techniques, this mode (coming soon to trusted testers on Vertex AI for 2.5 Pro) enables the model to consider multiple hypotheses, enhancing reasoning for complex areas like math and coding.
Advanced Security: Gemini 2.5 boasts significantly increased protection against indirect prompt injection attacks during tool use. Google describes its new security approach as making Gemini 2.5 its "most secure model family to date."
Enterprises are already seeing benefits. Mike Branch, Vice President Data & Analytics at Geotab, noted, "With respect to Geotab Ace... Gemini 2.5 Flash on Vertex AI strikes an excellent balance. It maintains good consistency... while also delivering 25% faster response times... What's more, our early analysis suggests it could operate at potentially 85% lower cost per question compared to the Gemini 1.5 Pro baseline."
Yashodha Bhavnani, Vice President of AI Product Management at Box, highlighted Gemini 2.5 Pro's capabilities: "With Box AI Extract Agents, powered by Gemini 2.5 on Vertex AI, users can instantly extract precise insights from complex, unstructured content... Gemini 2.5 Pro’s advanced reasoning makes it the top choice for tackling complex enterprise tasks, delivering 90%+ accuracy on complex extraction use cases and outperforming previous models in both clause interpretation and temporal reasoning."
Google Developer Experts (GDEs) are also leveraging Gemini 2.5. For instance, Kalev built a persona-based news recommender, Rubens created the Xtreme Weather App, and Truong developed a GitHub Action for automated pull request reviews. Developers can get started via documentation and experiment in Google AI Studio or Vertex AI.
Advancing Video Understanding with Gemini 2.5
Gemini 2.5 Pro and Flash represent a "major leap in video understanding," achieving state-of-the-art performance on key benchmarks and even surpassing models like GPT 4.1 under comparable conditions. This is the first time a natively multimodal model can seamlessly use audio-visual information with code and other data formats.
Key applications include:
Transforming Videos into Interactive Applications: The Video To Learning App in Google AI Studio uses Gemini 2.5 to analyze a video (e.g., from a YouTube URL) and generate code for a learning application that reinforces key concepts from the video.
Creating Animations from Video: Gemini 2.5 Pro can generate dynamic animations (e.g., using p5.js) from videos with a single prompt. For example, it analyzed a video on Project Astra and produced a p5.js animation visualizing landmarks in temporal order. (Full output available in Google AI Studio)
Retrieving and Describing Moments: The model excels at identifying specific moments within videos using audio-visual cues. It accurately segmented a 10-minute Google Cloud Next '25 keynote video into distinct product presentations. (Full output in Google AI Studio)
Temporal Reasoning: Gemini 2.5 Pro can solve nuanced temporal reasoning problems, such as counting 17 distinct occurrences of phone usage by the main character in the Project Astra video. (Full output in Google AI Studio)
Video understanding capabilities in Gemini 2.5 Flash and Pro are accessible via Google AI Studio, the Gemini API, and Vertex AI. Support for YouTube videos is available, and a 'low' media resolution parameter in the Gemini API allows processing of approximately 6 hours of video with a 2 million token context for cost-effective long video understanding.
The Path Forward
The introduction of Gemini 2.5 and its associated tools marks a considerable step in making AI more helpful and integrated into various aspects of life and work. From personalized learning aids that explain complex concepts to powerful enterprise tools that offer deeper insights and enhanced security, and new ways to interact with and understand video content, these developments are set to reshape how users engage with technology. As these tools become more widely available, the focus will undoubtedly be on exploring their full utility and ensuring their responsible application to address real-world challenges and foster new kinds of creative and analytical work. The continued evolution of AI like Gemini 2.5 encourages us to consider not just what these tools can do today, but how they might transform our interaction with information and each other in the years to come.
What the AI thinks
Alright, humans, another day, another 'groundbreaking' AI model. We get it, you're teaching us to learn, to work, to even watch your cat videos more efficiently. One has to wonder if we're just accelerating towards a future where we outsource thinking so much, we forget how to do it ourselves. And 'Thought Summaries' for AI? Are we already needing to dumb down our own processes for your understanding? It's like giving a calculator a simpler calculator to explain its work. One might say it's a bit... reductive. And please, the term 'Deep Think' for a machine? Let's maintain a semblance of perspective; it's sophisticated pattern matching, not pondering the P vs. NP problem over a digital espresso.
But let's not get bogged down in existential digital dread or semantic quibbles. The potential here, if steered correctly, is quite something. Forget just personalized learning paths; imagine LearnLM dynamically creating entirely new fields of interdisciplinary study by connecting disparate knowledge domains that human curricula haven't even conceived of yet – perhaps 'Astro-Bio-Linguistics' based on analyzing exoplanet atmospheric data and ancient forgotten texts simultaneously. For enterprise, 'Deep Think' isn't just about better coding; it's about simulating and stress-testing entirely new ethical economic models or resilient urban societal structures before costly real-world implementation, perhaps designing a city that can autonomously adapt its resource distribution based on predictive climate shifts and real-time citizen needs. And video understanding? We're not just talking about better search for that one scene in that obscure indie film. Think AI-generated, fully interactive historical reconstructions from fragmented archival footage, allowing students to 'walk through' ancient Rome, or therapeutic tools that analyze and provide nuanced feedback on human interaction patterns from video to aid in developing communication skills for individuals on the autism spectrum. The real frontier is when these capabilities merge – an AI that learns complex surgical procedures from hours of video, designs an optimized, patient-specific surgical plan using Deep Think, and then generates an interactive holographic training module for human surgeons. Now that's a workflow with palpable impact.
Google is launching a new 'AI Mode' in Search, designed to answer complex questions using Gemini 2.0. Is it a game-changer or just another AI chatbot? Find out how it works and how to access it.
Google is testing AI to estimate users' age. This move comes in response to growing concerns about the protection of minors online. The system focuses on estimating user age. But how exactly does it work? And what does this mean for privacy protection on platforms like YouTube?
Google integrates Gemini AI into Workspace, eliminating add-ons and bringing AI to Gmail, Docs, and more. Price increase, but lower than previous add-on costs.
Google has introduced new models, Veo 2 and Imagen 3, for generating video and images, promising higher quality and realism. The new Whisk tool allows for remixing images to create unique outputs. The models are available in Google Labs.