Gemini 1.5 Pro updates, 1.5 Flash arrival and 2 new Gemma models

Gemini 1.5 Pro updates, 1.5 Flash arrival and 2 new Gemma models

Last December, we unveiled Gemini 1. It did not work and the Gemini API was made available in Google AI Studio and Vertex AI for enterprise customers. In February we launched Gemini 1.5 Pro, the first model with a 1 million token context window is a pro. The other month, we introduced new features that are capable of understanding native audio, system instructions, JSON mode, and so on.

Besides, we started the production of Gemma, our family of lightweight open models that are built from the same research and technology that was used to create the Gemini models. Since February, the 2B and 7B models have been downloaded millions of times, and we added the variants last month for developers and researchers, for example, CodeGemma and RecurrentGemma.

now Gemini and Gemma are becoming even more. Today, we have the task of updating Gemini 1.5 Pro, introducing 1.5 Flash announcing the unveiling of new developer features, and accession of two new Gemma models.

Gemini 1. 5 Pro additions and a new 1. 5 Flash model

See also  Bard now helps with coding

Natively multimodal with long context: Both issue 1 and the significance of these works 1 have been discussed. 5 Pro and 1. 5 Flash have a 1 million token context window and thus you can interleave text, images, audio, and video as inputs. To get the benefits of 1, it is necessary to submit the form. 5 Pro has a 2 million token context window, therefore, you can join the waitlist in Google AI Studio or Vertex AI for Google Cloud customers.

The Gemini API’s price options and new developer features

  • New developer features: According to your suggestions, we are now introducing two new API features: video frame extraction and parallel function calling, which enable you to return more than one function call at a time. Finally, in June, we will implement context caching in Gemini 1. Thus, you only need to send the parts of your prompt, including large files, to the model once. This should make the long context even more useful and even more affordable.
  • Pricing: Moreover, we are also increasing the rate limits supported by our new pay-as-you-go service and providing access to the Gemini API for free in eligible regions through Google AI Studio. Look at the most recent prices for Google AI Studio and Vertex AI.

Enhancements to the Gemma family

  • PaliGemma: The first vision-language open model is now available and it is optimized for image captioning, visual Q&A, and other image labeling tasks. PaliGemma is combined with the other pre-trained Gemma variants, CodeGemma and RecurrentGemma.

  • Gemma 2: The next version of Gemma will be released in June and it is designed for industry-leading performance at the most useful developer sizes. Developers and researchers have demanded a bigger model that is still easy to use. The new Gemma 27B model should do that: It is more efficient than some models that are more than twice its size and will run on GPUs or a single TPU host in Vertex AI.

See also  Gemini sets new standards with a better model longer context AI agents and more

source

Leave a Comment

Comments

No comments yet. Why don’t you start the discussion?

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    This site uses Akismet to reduce spam. Learn how your comment data is processed.