Glamtush reports that Google DeepMind, the company’s AI unit, has announced Gemini 2.0, its next-generation AI model designed to power a new era of “agentic” experiences. This launch marks a significant leap forward in Google’s AI capabilities, as Gemini 2.0 boasts enhanced performance, multimodal understanding, and the ability to interact with the world in more sophisticated ways.
The first model in the Gemini 2.0 family, “Flash,” is now available as an experimental version for developers. It builds upon the success of its predecessor, Gemini 1.5 Flash, offering improved performance and faster response times, the company said.
Notably, Gemini 2.0 Flash outperforms even the more advanced Gemini 1.5 Pro on key benchmarks while operating at twice the speed, as per the company.
“In addition to supporting multimodal inputs like images, video, and audio, 2.0 Flash now supports multimodal output like natively generated images mixed with text and steerable text-to-speech (TTS) multilingual audio. It can also natively call tools like Google Search, code execution, as well as third-party user-defined functions,” said Demis Hassabis, CEO of Google DeepMind.
Google has listed new capabilities that are coming with the Gemini 2.0 Flash AI model, including:
Multimodal Input and Output: It can process and generate various types of data, including text, images, video, and audio.
Native Tool Use: It can integrate with tools like Google Search, code execution environments, and user-defined functions.
Enhanced Reasoning and Understanding: It demonstrates improved abilities in multimodal reasoning, long-context understanding, complex instruction following, and planning.
Developers can access Gemini 2.0 Flash through the Gemini API in Google AI Studio and Vertex AI. A new multimodal Live API is also being released to enable the creation of dynamic and interactive applications with real-time audio and video streaming capabilities.
Beyond developer tools, Gemini 2.0 Flash is being integrated into the Gemini app, Google’s AI assistant. Users can experience a chat-optimised version of the model, with plans to expand its availability to other Google products, such as Pixel smartphones, in the near future.
Google DeepMind is also exploring the frontiers of “agentic AI” with research prototypes like Project Astra, Project Mariner, and Jules, showcasing how Gemini 2.0 can enable AI agents to perform tasks, interact with users, and assist with complex activities like coding.
Tinubu is set to present the 2025 budget on Tuesday, December 17. Glamtush reports…
The Anambra State Executive Council has awarded more road contracts. Glamtush reports that the…
The Anambra State Executive Council has approved the medical expenses of a brutality victim.…
Selena Gomez has announced her engagement to Benny Blanco after one year of dating. …
Nigeria is set to add 150MW to the national grid before the end of the…
Podcasts are one of Africa’s favourite ways to tell stories. With almost 4 billion minutes…