Google’s Titans Just Solved AI’s Biggest Weakness, But...
AI Revolution
31,162 views • yesterday
Video Summary
Google has unleashed a torrent of AI updates, introducing a novel long-term memory system for extensive context windows and a more economical flash model that rivals its pro version. Simultaneously, they've been discreetly experimenting with AI-generated headlines and propelling Gemini's growth to outpace ChatGPT. Adding to this dynamic landscape, the Open AGI Foundation has unveiled Lux, a computer vision model poised to redefine AI agents. A particularly striking development is Google's "Titans" and "Myraz" research, which addresses transformer limitations with innovative memory architectures capable of handling over 2 million tokens and achieving remarkable accuracy on complex tasks, outperforming leading models.
One of the most significant revelations is the rapid growth of Gemini, which has seen a 30% increase in monthly active users in a recent period, significantly outpacing ChatGPT's slower growth. This surge is largely attributed to the popularity of the nano banana image generator and Gemini's integration within the Android ecosystem.
Short Highlights
- Google's advancements include a new long-term memory system for massive context windows and a cost-effective "nano banana 2 flash" model.
- The Open AGI Foundation released Lux, a computer vision model designed to interact directly with user interfaces, outperforming competitors on benchmarks like Mind 2 Web with an 83.6 score.
- Google's Gemini has seen a significant growth spurt, with monthly active users rising 30% in a recent period, far outpacing ChatGPT's 6% growth.
- Google quietly tested AI-generated headlines on discover users, leading to significant misinterpretations and concerns about trust from newsrooms.
- Gemini's user engagement has surged by 120% since March, with users spending 11 minutes per day in the app, largely driven by the popular nano banana image generator.
Key Details
Google's AI Innovations: Titans and Myraz [00:02]
- Google is significantly advancing its AI ecosystem with updates to long-term memory systems, introducing a new "nano banana 2 flash" model, experimenting with AI-generated headlines, and boosting Gemini's growth.
- The "Titans" research aims to overcome transformer limitations with long inputs by blending short-term memory (window attention) and a dynamic long-term memory module that updates during use.
- This long-term memory system stores information based on surprise and forgets in a smart way, building three types: memory as context, memory as gate, and memory as layer.
- The "Mac version" of Titans has demonstrated exceptional performance, handling context windows over 2 million tokens and outperforming models like GPT-4 and Llama 3.170B on benchmarks like Babylon.
- "Myraz" is presented as a framework for understanding sequence models, breaking down memory into shape, storage decision, eviction rate, and update mechanisms, leading to new attention-free models like Moneta, Yad, and Mamora.
"Titans is Google trying to blend both worlds."
Lux: The Next-Generation AI Agent [04:24]
- The Open AGI Foundation has released Lux, a computer vision model that operates directly on user interfaces, outputting clicks, scrolls, and key presses, rather than relying solely on APIs.
- Lux excels at handling real-world UI interactions across desktops, browsers, and applications, achieving an 83.6 score on the Mind 2 Web benchmark, significantly higher than competitors like Gemini CUA (69.0) and OpenAI Operator (61.3).
- It features three levels of autonomy: Actor mode for straightforward tasks, Thinker mode for broader goals, and Tasker mode for deterministic execution with error handling.
- Lux's training utilizes "agentic active pre-training," allowing it to learn through direct interaction within digital environments, leading to a more intuitive understanding of real interfaces.
- The model is also reportedly 10 times cheaper per token than OpenAI's Operator, making long, multi-step automation more economically viable.
"Lux isn't a chat model with some browser plugin slapped onto it. It actually looks at the screen, reads the UI, and outputs clicks, scrolls, key presses, everything."
Nano Banana 2 Flash and AI Headline Experiments [07:11]
- Google is preparing to release "Nano Banana 2 Flash," a more cost-effective version of its image generation model that offers performance comparable to the Pro version.
- This strategy allows Google to deploy flash models for high-volume scenarios where cost efficiency is paramount, especially for users engaging in high-frequency image generation or large batch workloads.
- In a separate, controversial experiment, Google Discover was observed rewriting headlines for Android users, sometimes drastically altering the original meaning and causing concern among newsrooms about trust erosion.
- Google stated this was a "small UI experiment" to make topics easier to scan, but did not fully address why many rewritten titles deviated significantly from the articles' content.
"The public announcement is probably coming sometime in December."
Gemini's Rapid Growth and Competitive Landscape [10:13]
- New user data reveals Gemini is experiencing substantial growth, with monthly active users (MAUs) rising 30% in a recent period, significantly outperforming ChatGPT's 6% growth during the same timeframe.
- While ChatGPT still leads in global downloads (around 50%) and MAUs (around 55% or 810 million users), its growth has slowed considerably compared to Gemini's rapid ascent.
- Gemini's MAU share has climbed by 3 percentage points, while ChatGPT's has dropped by 3 percentage points, indicating a shift in market momentum.
- The surge in Gemini's popularity is largely driven by the "nano banana" image generator and its integration within the Android ecosystem, making it the default AI for many users.
- In response to Google's acceleration, OpenAI is prioritizing upgrades and has put advertising experiments on hold, focusing on developing a new model named "Garlic" to compete with Gemini 3.
"The timing shows how seriously they're taking Google's acceleration."
Other People Also See