A glimpse into Gemini, Google’s most advanced AI model to date

Google just rolled out a new AI powerhouse named Gemini, which it says is its most impressive AI model to date.

A glimpse into Gemini, Google’s most advanced AI model to date
Source: Google

The backstory: Last November, OpenAI dropped ChatGPT, the artificial intelligence (AI) chatbot that quickly became a global sensation. A few months later, Microsoft’s Bing joined the AI race with its own chatbot to amp up its search engine game. Google’s Bard is also in the mix but hasn’t grabbed as much attention. It got some upgrades, like accessing data across Google products, but it hasn’t quite matched up to heavy hitters like ChatGPT.

More recently: In August, Google started trying out something new called the Search Generative Experience (SGE), which is an AI-powered search product that aims to make online searching both more conversational and more comprehensive. But it’s still in the testing phase. There's no confirmed date for when it might be available to everyone. 

The development: Google just rolled out a new AI powerhouse named Gemini, which it says is its most impressive AI model to date. Gemini comes in three versions – Ultra (the biggest, most capable model), Pro (an all-purpose model that covers a wide range of tasks) and Nano (which is more compact for specific jobs and mobile use). Google plans to license the model through Google Cloud so developers can put it in their apps.

From December 13 onward, developers and enterprise customers can get their hands on Gemini Pro via the Gemini API in Google AI Studio or Cloud Vertex AI. Android developers will also be able to tap into the Nano version. Google plans to use Gemini in its own products like Bard and SGE. Gemini Ultra is the first model to outperform humans in specific kinds of testing, and execs said the Pro version has even outperformed ChatGPT’s GPT-3.5. They did avoid answering questions about how it compares to OpenAI’s GPT-4, though. 

Key comments: 

“Gemini is the result of large-scale collaborative efforts by teams across Google, including our colleagues at Google Research,” wrote CEO Sundar Pichai in a blog post. “It was built from the ground up to be multimodal, which means it can generalize and seamlessly understand, operate across and combine different types of information including text, code, audio, image and video.”

“This new era of models represents one of the biggest science and engineering efforts we’ve undertaken as a company,” said Pichai in a blog post. “I’m genuinely excited for what’s ahead and for the opportunities Gemini will unlock for people everywhere.”

“We are continuing to invest in the teams that work on applying our AI Principles to our technology,” said Brian Gabriel, a spokesperson for Google, to Bloomberg in April.