A note from Sundar Pichai, CEO of Google and Alphabet:
Information is at the heart of human progress. That’s why we’ve focused for more than 26 years on our mission to organize the world’s information and make it accessible and useful. And that’s why we continue to push the boundaries of AI to organize this information on every input and make it accessible through any output, so that it can be truly useful to you.
This was our vision when we introduced Gemini 1.0 last December. The first model designed to be natively multimodal, Gemini 1.0 and 1.5 brought great advances in multimodality and long context to understand information contained in text, video, images, audio and code, and in process much more.
Today, millions of developers build with Gemini. And it helps us reinvent all our products – including the 7 with 2 billion users – and create new ones. CarnetLM is a great example of what multimodality and long context can enable people, and why it is loved by so many.
Over the past year, we’ve invested in developing more agent-based models, meaning they can better understand the world around you, think across multiple stages, and act on your behalf, under your supervision.
Today, we’re excited to launch our next generation of models built for this new agentic era: we’re introducing Gemini 2.0, our most capable model yet. With new advances in multimodality – such as native image and audio output – and the use of native tools, this will allow us to create new AI agents that will bring us closer to our vision of a universal assistant.
Today we are putting version 2.0 into the hands of trusted developers and testers. And we’re working quickly to integrate it into our products, leading with Gemini and Search. Starting today, our experimental Gemini 2.0 Flash model will be available to all Gemini users. We are also launching a new feature called In-depth researchwho uses advanced reasoning and long context capabilities to act as a research assistant, exploring complex topics and compiling reports on your behalf. It is available in Gemini Advanced today.
No product has been transformed by AI as much as search. Our AI insights now reach 1 billion people, allowing them to ask entirely new types of questions, quickly becoming one of our most popular search features. In the next step, we integrate Gemini 2.0’s advanced reasoning capabilities with AI insights to tackle more complex topics and multi-step questions, including advanced math equations, multimodal queries, and coding. We began limited testing this week and will roll it out more widely early next year. And we will continue to offer AI previews in more countries and languages over the next year.
Advances in 2.0 are supported by decades-long investments in our differentiated, full-stack approach to AI innovation. It’s built on custom hardware like Trillium, our sixth generation TPUs. TPUs powered 100% of Gemini 2.0 training and inference, and today Trillium is generally available to customers so they can build with it too.
If Gemini 1.0 was about organizing and understanding information, Gemini 2.0 is about making it much more useful. I can’t wait to see what this next era brings us.
-Sundar