Gemma making waves!

By Martin Thomas, 22 February, 2024

It all started a little way back with LLaMA and then the outstanding LLaMA 2 the Large Language Model from Google that was released as a platform in open source for others to build on.

Now Google is well and truly back in the opensource LLM world with its new model Gemma!

Gemma has been trained on 6 million tokens compared to LLaMA 2's 4 million.

Gemma has been released with 4 new models.

The Base Models

Firstly a base model with a 7 billion or 2 billion parameters model. variants

This offers raw text processing and all-purpose general capabilities.

The Instruction-Tuned Models

Secondly the 'instruction-tuned' model which also has 7 billion and 2 billion parameters model variants.

This offers the ability to specifically follow instructions and get tasks done effectively.

Google describes these models and versions as a 'family' of lightweight, opensource models built from the Gemini models.

Advantages

So what does this all mean?  

Firstly Google says they are light-weight meaning they can run on your own machine, your laptop or on a cloud platform.

Secondly, they are 'open-source' meaning they are available free for development.

Thirdly they are performing well in tests - see sites like Hugging Face.

Next they are pre-trained and tuned meaning different versions can address very specific needs.  Customisation can be further addressed with tools like LoRA.

On the downside for the moment

The models are not multimodal meaning they deal with text-in and text-out tasks and currently only operate in the English language.

Google will no doubt continue to improve and go for the next version but this is a big step forward.

 

If you would like us to help you with your digital solutions journey then contact us!

 

Blog tags
Blog Image
Google's Gemma
Blog short description
Google's new open-source LLM