Be part of our day by day and weekly newsletters for the most recent updates and unique content material on industry-leading AI protection. Study Extra
Google’s Gemini sequence of AI giant language fashions (LLMs) began off tough practically a yr in the past with some embarrassing incidents of picture technology gone awry, nevertheless it has steadily improved since then, and the corporate seems to be intent on making its second-generation effort — Gemini 2.0 — the largest and finest but for shoppers and enterprises.
In the present day, the corporate introduced the final launch of Gemini 2.0 Flash, launched Gemini 2.0 Flash-Lite, and rolled out an experimental model of Gemini 2.0 Professional.
These fashions, designed to assist builders and companies, at the moment are accessible via Google AI Studio and Vertex AI, with Flash-Lite in public preview and Professional accessible for early testing.
“All of those fashions will characteristic multimodal enter with textual content output on launch, with extra modalities prepared for basic availability within the coming months,” Koray Kavukcuoglu, CTO of Google DeepMind, wrote within the firm’s announcement weblog publish — showcasing a bonus Google is bringing to the desk whilst opponents equivalent to DeepSeek and OpenAI proceed to launch highly effective rivals.
Google performs to its multimodal strenghts
Neither DeepSeek-R1 nor OpenAI’s new o3-mini mannequin can settle for multimodal inputs — that’s, pictures and file uploads or attachments.
Whereas R1 can settle for them on its web site and cell app chat, The mannequin performs optical character recognition (OCR) a greater than 60-year-old know-how, to extract the textual content solely from these uploads — not truly understanding or analyzing any of the opposite options contained therein.
Nevertheless, each are a brand new class of “reasoning” fashions that intentionally take extra time to assume via solutions and replicate on “chains-of-thought” and the correctness of their responses. That’s against typical LLMs just like the Gemini 2.0 professional sequence, so the comparability between Gemini 2.0, DeepSeek-R1 and OpenAI o3 is a little bit of an apples-to-oranges.
However there was some information on the reasoning entrance right this moment from Google, too: Google CEO Sundar Pichai took to the social community X to declare that the Google Gemini cell app for iOS and Android has been up to date with Google’s personal rival reasoning mannequin Gemini 2.0 Flash Considering. The mannequin may be linked to Google Maps, YouTube and Google Search, permitting for an entire new vary of AI-powered analysis and interactions that merely can’t be matched by upstarts with out such providers like DeepSeek and OpenAI.
I attempted it briefly on the Google Gemini iOS app on my iPhone whereas scripting this piece, and it was spectacular based mostly on my preliminary queries, pondering via the commonalities of the highest 10 hottest YouTube movies of the final month and likewise offering me a desk of close by docs’ places of work and opening/closing hours, all inside seconds.
![](https://venturebeat.com/wp-content/uploads/2025/02/IMG_1568.png?w=276)
![](https://venturebeat.com/wp-content/uploads/2025/02/IMG_1571.png?w=276)
![](https://venturebeat.com/wp-content/uploads/2025/02/IMG_1575_e5459f.png?w=276)
Gemini 2.0 Flash enters basic launch
The Gemini 2.0 Flash mannequin, initially launched as an experimental model in December, is now production-ready.
Designed for high-efficiency AI functions, it gives low-latency responses and helps large-scale multimodal reasoning.
One main profit over the competitors is in its context window, or the variety of tokens that the consumer can add within the type of a immediate and obtain again in a single back-and-forth interplay with an LLM-powered chatbot or utility programming interface (API).
Whereas many main fashions, equivalent to OpenAI’s new o3-mini that debuted final week, solely assist 200,000 or fewer tokens — in regards to the equal of a 400 to 500 web page novel — Gemini 2.0 Flash helps 1 million, that means it’s is able to dealing with huge quantities of knowledge, making it notably helpful for high-frequency and large-scale duties.
Gemini 2.0 Flash-Lite arrives to bend the fee curve to the bottom but
Gemini 2.0 Flash-Lite, in the meantime, is an all-new LLM geared toward offering an economical AI answer with out compromising on high quality.
Google DeepMind states that Flash-Lite outperforms its full-size (bigger parameter-count) predecessor, Gemini 1.5 Flash, on third-party benchmarks equivalent to MMLU Professional (77.6% vs. 67.3%) and Chicken SQL programming (57.4% vs. 45.6%), whereas sustaining the identical pricing and pace.
It additionally helps multimodal enter and includes a context window of 1 million tokens, just like the total Flash mannequin.
Presently, Flash-Lite is accessible in public preview via Google AI Studio and Vertex AI, with basic availability anticipated within the coming weeks.
As proven within the desk under, Gemini 2.0 Flash-Lite is priced at $0.075 per million tokens (enter) and $0.30 per million tokens (output). Flash-Lite is positioned as a extremely reasonably priced choice for builders, outperforming Gemini 1.5 Flash throughout most benchmarks whereas sustaining the identical value construction.
![](https://venturebeat.com/wp-content/uploads/2025/02/GjCemCUW4AEUCHT.jpg?w=800)
Logan Kilpatrick highlighted the affordability and worth of the fashions, stating on X: “Gemini 2.0 Flash is the perfect worth prop of any LLM, it’s time to construct!”
Certainly, in comparison with different main conventional LLMs accessible through supplier API, equivalent to OpenAI 4o-mini ($0.15/$0.6 per 1 million tokens in/out), Anthropic Claude ($0.8/$4! per 1M in/out) and even DeepSeek’s conventional LLM V3 ($0.14/$0.28), in Gemini 2.0 Flash seems to be the perfect bang for the buck.
Gemini 2.0 Professional arrives in experimental availability with 2-million token context window
For customers requiring extra superior AI capabilities, the Gemini 2.0 Professional (experimental) mannequin is now accessible for testing.
Google DeepMind describes this as its strongest mannequin for coding efficiency and the flexibility to deal with advanced prompts. It includes a 2 million-token context window and improved reasoning capabilities, with the flexibility to combine exterior instruments like Google Search and code execution.
Sam Witteveen, co-founder and CEO of Purple Dragon AI and an exterior Google developer professional for machine studying who usually companions with VentureBeat, mentioned the Professional mannequin in a YouTube assessment. “The brand new Gemini 2.0 Professional mannequin has a two-million-token context window, helps instruments, code execution, perform calling and grounding with Google Search — every little thing we had in Professional 1.5, however improved.”
He additionally famous of Google’s iterative strategy to AI improvement: “One of many key variations in Google’s technique is that they launch experimental variations of fashions earlier than they go GA (usually accessible), permitting for speedy iteration based mostly on suggestions.”
Efficiency benchmarks additional illustrate the capabilities of the Gemini 2.0 mannequin household. Gemini 2.0 Professional, as an example, outperforms Flash and Flash-Lite throughout duties like reasoning, multilingual understanding and long-context processing.
![](https://venturebeat.com/wp-content/uploads/2025/02/image6.original_JY99INi-1.png?w=512)
AI security and future developments
Alongside these updates, Google DeepMind is implementing new security and safety measures for its Gemini 2.0 fashions. The corporate is leveraging reinforcement studying strategies to enhance response accuracy, utilizing AI to critique and refine its personal outputs. Moreover, automated safety testing is getting used to establish vulnerabilities, together with oblique immediate injection threats.
Wanting forward, Google DeepMind plans to broaden the capabilities of the Gemini 2.0 mannequin household, with further modalities past textual content anticipated to turn out to be usually accessible within the coming months.
With these updates, Google is reinforcing its push into AI improvement, providing a variety of fashions designed for effectivity, affordability and superior problem-solving, and answering the rise of DeepSeek with its personal suite of fashions starting from highly effective to very highly effective and intensely reasonably priced to barely much less (however nonetheless significantly) reasonably priced.
Will or not it’s sufficient to assist Google eat into among the enterprise AI market, which was as soon as dominated by OpenAI and has now been upended by DeepSeek? We’ll hold monitoring and allow you to know!