Gemini 2.0 is now available to everyone
Google has launched Gemini 2.0, featuring the Flash model for all users, a Pro model for coding, and a cost-efficient Flash-Lite model, all with enhanced safety measures and ongoing updates.
Read original articleGoogle has announced the general availability of its Gemini 2.0 model, which includes several updates aimed at enhancing performance and usability for developers and users. The Gemini 2.0 Flash model, known for its efficiency and low latency, is now accessible to all users via the Gemini app and API in Google AI Studio and Vertex AI. This model supports multimodal input and has a context window of 1 million tokens. Additionally, an experimental version of Gemini 2.0 Pro has been released, which is optimized for coding tasks and complex prompts, featuring a larger context window of 2 million tokens. The new Gemini 2.0 Flash-Lite model is also introduced, designed to be cost-efficient while maintaining quality. All models are built with safety measures, including reinforcement learning techniques to improve response accuracy and security against potential cyber threats. Google emphasizes its commitment to ongoing improvements and updates for the Gemini 2.0 family of models.
- Gemini 2.0 Flash is now generally available for all users and developers.
- The Gemini 2.0 Pro model is optimized for coding and complex prompts, featuring a 2 million token context window.
- Gemini 2.0 Flash-Lite is introduced as a cost-efficient model with improved quality.
- All models incorporate safety measures to enhance response accuracy and mitigate security risks.
- Google plans to continue updating and improving the Gemini 2.0 models.
Related
Gemini Pro 1.5 experimental "version 0801" available for early testing
Google DeepMind's Gemini family of AI models, particularly Gemini 1.5 Pro, excels in multimodal understanding and complex tasks, featuring a two million token context window and improved performance in various benchmarks.
Google Gemini 1.5 Pro leaps ahead in AI race, challenging GPT-4o
Google has launched Gemini 1.5 Pro, an advanced AI model excelling in multilingual tasks and coding, now available for testing. It raises concerns about AI safety and ethical use.
Two new Gemini models, reduced 1.5 Pro pricing, increased rate limits, and more
Google updated its Gemini models, introducing Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002, featuring over 50% price reduction, increased rate limits, improved performance, and free access for developers via Google AI Studio.
Gemini 1.5 Flash-8B is now production ready
Google announced the production-ready Gemini 1.5 Flash-8B, featuring a 50% price reduction, double rate limits, and optimized performance for high-volume tasks, accessible for free via Google AI Studio.
Gemini 2.0: our new AI model for the agentic era
Google has launched Gemini 2.0, an advanced AI model with multimodal capabilities, including image and audio output. Wider access is planned for early 2025, focusing on responsible AI development.
- Users express confusion over the various models and their availability, particularly regarding the "Gemini Advanced" and subscription details.
- Many commenters highlight the competitive pricing of Gemini 2.0 models compared to other AI offerings, noting cost-effectiveness.
- There are mixed reviews on the performance of Gemini 2.0, with some praising its capabilities while others criticize its limitations and quality compared to competitors.
- Concerns about the naming conventions of the models and the clarity of their functionalities are frequently mentioned.
- Users are eager for more information on features like coding performance and video processing capabilities.
I find the lack of clarity very frustrating. If I want to try Google's "best" model, should I be purchasing something? AI Studio seems focused around building an LLM wrapper app, but I just want something to answer my questions.
Edit: what I've learned through Googling: (1) if you search "is gemini advanced included with workspace" you get an AI overview answer that seems to be incorrect, since they now include Gemini Advanced (?) with every workspace subscription.(2) a page exists telling you to buy the add-on (Gemini for Google Workspace), but clicking on it says this is no longer available because of the above. (3) gemini.google.com says "Gemini Advanced" (no idea which model) at the top, but gemini.google.com/advanced redirects me to what I have deduced is the consumer site (?) which tells me that Gemini Advanced is another $20/month
The problem, Google PMs if you're reading this, is that the gemini.google.com page does not have ANY information about what is going on. What model is this? What are the limits? Do I get access to "Deep Research"? Does this subscription give me something in aistudio? What about code artifacts? The settings option tells me I can change to dark mode (thanks!).
Edit 2: I decided to use aistudio.google.com since it has a dropdown for me on my workspace plan.
> Gemini 2.0, 2.0 Pro and 2.0 Pro Experimental, Gemini 2.0 Flash, Gemini 2.0 Flash Lite
3 different ways of accessing the API, more than 5 different but extremely similarly named models. Benchmarks only comparing to their own models.
Can't be more "Googley"!
We started talking about my plans for the day, and I said I was making chili. G asked if I have a recipe or if I needed one. I said, I started with Obama's recipe many years ago and have worked on it from there.
G gave me a form response that it can't talk politics.
Oh, I'm not talking politics, I'm talking chili.
G then repeated form response and tried to change conversation, and as long as I didn't use the O word, we were allowed to proceed. Phew
Sonnet 3.5 v2
o3-mini-high
Gemini Flash-Lite
It's like a competition to see who can make the goofiest naming conventions.
Regarding model quality, we experiment with Google models constantly at Rev and they are consistently the worst of all the major players. They always benchmark well and consistently fail in real tasks. If this is just a small update to the gemini-exp-1206 model, then I think they will still be in last place.
It's a great way to experiment with all the Gemini models that are also available via the API.
If you haven't yet, try also Live mode at https://aistudio.google.com/live.
You can have a live conversation with Gemini and have the model see the world via your phone camera (or see your desktop via screenshare on the web), and talk about it. It's quite a cool experience! It made me feel the joy of programming and using computers that I had had so many times before.
The pricing is interesting: Gemini 2.0 Flash-Lite is 7.5c/million input tokens and 30c/million output tokens - half the price of OpenAI's GPT-4o mini (15c/60c).
Gemini 2.0 Flash isn't much more: 10c/million for text/image input, 70c/million for audio input, 40c/million for output. Again, cheaper than GPT-4o mini.
In my experience, I'd reach for Gemini 2.0 Flash over 4o in a lot of multimodal/document use cases. Especially given the differences in price ($0.10/million input and $0.40/million output versus $2.50/million input and $10.00/million output).
That being said, Qwen2.5 VL 72B and 7B seem even better at document image tasks and localization.
[1] https://notes.penpusher.app/Misc/Google+Gemini+101+-+Object+...
I highly recommend using it via https://aistudio.google.com/. Gemini app has some additional bells and whistles, but for some reason quality isn't always on par with aistudio. Also Gemini app seems to have more filters -- it seems more shy answering controversial topics. Just some general impressions.
> Today, we’re releasing an experimental version of Gemini 2.0 Pro that responds to that feedback. It has the strongest coding performance and ability to handle complex prompts, with better understanding and reasoning of world knowledge, than any model we’ve released so far. It comes with our largest context window at 2 million tokens, which enables it to comprehensively analyze and understand vast amounts of information, as well as the ability to call tools like Google Search and code execution.
Audio input is $0.70 per million tokens on 2.0 Flash, $0.075 for 2.0 Flash-Lite and 1.5 Flash.
For gpt-4o-mini-audio-preview, it's $10 per million tokens of audio input.
Next release should be called Gemini Macromedia
My experience with the Gemini 1.5 models has been positive. I think Google has caught up.
1.5 pro and the old 2.0 flash experimental generated responses in AI studio but the new 2.0 models respond with blank answers.
I wonder if it's timing out or some sort of newer censorship models is preventing 2.0 from answering my query. The novel is pg-13 at most but references to "bronze skinned southern barbarians" "courtesans" "drugs" "demonic sects" and murder could I guess set it off.
The 2 million token window sure feels exciting.
Their search costs 7x Perplexity Sonar's but imagine a lot of people will start with Google given they can get a pretty decent amount of search for free now.
As in I have a video file I want to send it to the model and get a response about it. Not their 'live stream' or whatever functionality.
How many tokens can gemini.google.com handle as input? How large is the context window before it forgets? A quick search said it's 128k token window but that applies to Gemini 1.5 Pro, how is it now then?
My assumption is that "Gemini 2.0 Flash Thinking Experimental is just" "Gemini 2.0 Flash" with reasoning and "Gemini 2.0 Flash Thinking Experimental with apps" is just "Gemini 2.0 Flash Thinking Experimental" with access to the web and Googles other services, right? So sticking to "Gemini 2.0 Flash Thinking Experimental with apps" should be the optimal choice.
Is there any reason why Gemini 1.5 Flash is still an option? Feels like it should be removed as an option unless it does something better than the other.
I have difficulties understanding where each variant of the Gemini model is suited the most. Looking at aistudio.google.com, they have already update the available models.
Is "Gemini 2.0 Flash Thinking Experimental" on gemini.google.com just "Gemini experiment 1206" or was it "Gemini Flash Thinking Experimental" aistudio.google.com?
I have a note on my notes app where I rank every llm based on instructions following and math, to this day, I've had difficulties knowing where to place every Gemini model. I know there is a little popup when you hover over each model that tries to explain what each model does and which tasks it is best suited for, but these explanations have been very vague to me. And I haven't even started on the Gemini Advanced series or whatever I should call it.
The available models on aistudio is now:
- Gemini 2.0 Flash (gemini-2.0-flash)
- Gemini 2.0 Flash Lite Preview (gemini-2.0-flash-lite-preview-02-05)
- Gemini 2.0 Pro Experimental (gemini-2.0-pro-exp-02-05)
- Gemini 2.0 Flash Thinking Experimental (gemini-2.0-flash-thinking-exp-01-21)
If I had to sort these from most likely to fulfill my need to least likely, then it would probably be:
gemini-2.0-flash-thinking-exp-01-21 > gemini-2.0-pro-exp-02-05 > gemini-2.0-flash-lite-preview-02-05 > gemini-2.0-flash
Why? Because aistudio describes gemini-2.0-flash-thinking-exp-01-21 as being able to tackle most complex and difficult tasks while gemini-2.0-pro-exp-02-05 and gemini-2.0-flash-lite-preview-02-05 only differs with how much context they can handle.
So with that out of the way, how does Gemini-2.0-flash-thinking-exp-01-21 compare against o3-mini, Qwen 2.5 Max, Kimi k1.5, DeepSeek R1, DeepSeek V3 and Sonnet 3.5?
My current list of benchmarks I go through is artificialanalysis.ai, lmarena.ai, livebench.ai and aider.chat:s polygot benchmark but still, the whole Gemini suite is difficult to reason and sort out.
I feel like this trend of having many different models with the same name but different suffix starts be an obstacle to my mental model.
stumped it
Related
Gemini Pro 1.5 experimental "version 0801" available for early testing
Google DeepMind's Gemini family of AI models, particularly Gemini 1.5 Pro, excels in multimodal understanding and complex tasks, featuring a two million token context window and improved performance in various benchmarks.
Google Gemini 1.5 Pro leaps ahead in AI race, challenging GPT-4o
Google has launched Gemini 1.5 Pro, an advanced AI model excelling in multilingual tasks and coding, now available for testing. It raises concerns about AI safety and ethical use.
Two new Gemini models, reduced 1.5 Pro pricing, increased rate limits, and more
Google updated its Gemini models, introducing Gemini-1.5-Pro-002 and Gemini-1.5-Flash-002, featuring over 50% price reduction, increased rate limits, improved performance, and free access for developers via Google AI Studio.
Gemini 1.5 Flash-8B is now production ready
Google announced the production-ready Gemini 1.5 Flash-8B, featuring a 50% price reduction, double rate limits, and optimized performance for high-volume tasks, accessible for free via Google AI Studio.
Gemini 2.0: our new AI model for the agentic era
Google has launched Gemini 2.0, an advanced AI model with multimodal capabilities, including image and audio output. Wider access is planned for early 2025, focusing on responsible AI development.