Google launches its open-source model Gemma 2!
Hi AI Enthusiasts,
Welcome to this week's Magic AI News, where we present you the most exciting AI and tech news of the week. We organize the updates for you and share our thoughts with you!
This week's Magic AI tool is a must-know tool for all podcast fans! Stay curious! ๐
Let's explore this week's AI news together. ๐๐ฝ
In today's Magic AI:
Google launches its open-source model Gemma 2
OpenAI presents CriticGPT to improve ChatGPT's code output
Efficient AI language models without matrix multiplication
Magic AI tool of the week
Hand-picked articles of the week
๐ค Check out our hand-picked AI tool list
Top AI news of the week
๐ญ Google launches its open-source model Gemma 2
Google Gemma 2 is now available for developers. Gemma is a family of lightweight, state-of-the-art open models based on the same technology used to create Google's Gemini models.
Gemma 2 comes in two sizes with 9B and 27B parameters. In addition, Google announced a 2.6B parameter model. The 27B model offers competitive performance compared to models that are more than twice its size. According to Google, the model 9B outperforms similar models such as the Llama 3 8B.
Benchmark of Gemma 2 (Image by Google)
Our thoughts
Google has launched powerful and competitive open-source LLMs. In addition, it can run on a single NVIDIA H100 Tensor Core GPU or TPU host, significantly reducing deployment costs. We think the model will enjoy great popularity in the open-source community. The model is available on HuggingFace.
More information
- Gemma 2 is now available to researchers and developers - Google Blog
๐ค OpenAI presents CriticGPT to improve ChatGPT's code output
OpenAI has presented an GPT-4 based AI model, called CriticGPT, that is designed to find errors in ChatGPT's code output. According to OpenAI's blog post, people who use both for coding perform 60 % better than people who only use ChatGPT. CriticGPT makes suggestions when errors are detected and asks the user for feedback. OpenAI uses the input to improve ChatGPT.
CriticGPT does not solve the problem of hallucinations, but it should lead to more precise answers. OpenAI named some limitations that the model currently still has. They have trained CriticGPT on short ChatGPT responses. In the future, more complex responses need to be included in the training. In addition, errors in the answers can be very complicated, and these must also be considered in the future. Currently, CriticGPT can only help to a limited level.
Our thoughts
Nowadays, LLMs still make some mistakes, which are known as hallucinations. For this reason, LLMs should not be used in critical areas. We welcome the current progress made by OpenAI, but there is still a long way to go before LLMs work almost perfectly. We will follow the research on this topic further.
More information
Finding GPT-4โs mistakes with GPT-4 - OpenAI Blog
LLM Critics Help Catch LLM Bugs - OpenAI paper
๐ค Efficient AI language models without matrix multiplication
Researchers from the USA and China have developed language models that don't need memory-intensive matrix multiplications. According to the paper, the models can compete with modern transformers. Matrix multiplications are responsible for most of the resource requirements and influence the scaling of the models.
Furthermore, the authors provide a GPU-efficient implementation of this proposed MatMul-free model that reduces memory consumption during training by up to 61% compared to a non-optimized baseline solution. By using an optimized kernel during inference, the memory consumption of the model can be reduced by more than ten times compared to non-optimized models.
For more in-depth information, we recommend reading the full paper. In addition, the authors provide the code implementation on GitHub.
Our thoughts
Matrix multiplications (MatMul) are responsible for much of the overall computational cost of large language models (LLMs). The training of LLMs needs a huge amount of energy. For this reason, it is important to find more efficient methods so that AI progress and climate protection can go hand-in-hand.
More information
MatMul-Free LM - GitHub repo
Magic AI tool of the week
๐ชSnipd - Unleash the Knowledge in your Podcasts
Snipd allows you to mark highlights in a podcast. This way, you can access the most valuable information from the podcast later. Currently, the tool supports 11 languages (English, German, Spanish, French, ...) and more are planned.
If you love podcasts, then try it out!
Step-by-Step Guide
You hear something interesting.
Tap your headphones to save it to your library.
An AI generates transcripts, summaries, and titles.
Export your saved highlights to your favorite app like Notion*.
Articles of the week
๐ You might also be interested in these blog articles:
Mistral's Codestral: Create a local AI Coding Assistant for VSCode
Responsible Development of an LLM Application + Best Practices
Thanks for reading, and see you next time.
- Tinz Twins
P.S. Have a nice weekend! ๐๐
* Disclosure: The links are affiliate links, which means we will receive a commission if you purchase through these links. There are no additional costs for you.
Subscribe to my newsletter
Read articles from Tinz Twins directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
Tinz Twins
Tinz Twins
Hey, we are the Tinz Twins! ๐๐ฝ ๐๐ฝ We both have a Bachelor's degree in Computer Science and a Master's degree in Data Science. In our blog articles, we deal with topics around Data Science.