Google’s Gemma 2: Free AI Beats 70B Models at 27B Size

Google has been making significant strides in the open source AI landscape, rapidly expanding its Gemma 2 series of lightweight, state-of-the-art models. The tech giant’s commitment to democratizing AI is evident in its latest releases, which have positioned Google as a formidable rival to Meta in the open source AI race.

Gemma 2: Powerful Performance in Compact Packages

The Gemma 2 series, built using the same cutting-edge research and technology as Google’s Gemini models, offers best-in-class performance for its size. Notably, Gemma models outperform significantly larger models on key benchmarks while adhering to rigorous standards for safe and responsible outputs.

Key highlights of the Gemma 2 series include:

  • Gemma 2B: A compact powerhouse with 2.6B parameters, trained on 20 trillion tokens, and capable of running on mobile devices.
  • Gemma 9B: Outperforming comparable models like LLaMa 3 8B, with robust performance from training on 8 trillion tokens.
  • Gemma 27B: Matching the quality of larger models such as Mixtral 8x22B, trained on an impressive 13 trillion tokens.

Unsloth Team Delivers Rapid Analysis of Gemma 2 2B

True to form, the Unsloth team wasted no time in providing a comprehensive overview of the Gemma 2 2B model. Key highlights include:

  • Impressive performance on the LYMSYS arena, surpassing GPT 3.5 and Mixtral 8x7B
  • Strong showing on the HF v2 leaderboard, particularly in IFEval, despite struggling in MATH
  • MMLU score of 56.1 and MBPP score of 36.6
  • Outperforming the previous Gemma v1 2B by over 10% in benchmarks
  • 2.6B parameters and multilingual capabilities
  • Trained on 20 trillion tokens
  • Potentially distilled from Gemma 2 27B (unconfirmed)
  • Trained on 512 TPU v5e

The model is now available on Hugging Face: https://huggingface.co/google/gemma-2-2b

Responsible AI at the Forefront

Google’s commitment to responsible AI shines through in the Gemma series. Extensive measures have been taken to filter out sensitive data from training sets, align models with responsible behaviors through reinforcement learning, and conduct robust evaluations to understand and reduce risk profiles.

The release of the Responsible Generative AI Toolkit alongside Gemma further underscores Google’s dedication to promoting safe and responsible AI development.

Optimized for Diverse Frameworks and Hardware

Gemma’s versatility is evident in its compatibility with a wide range of frameworks, devices, and platforms. From multi-framework tools like Keras 3.0, PyTorch, JAX, and Hugging Face Transformers to cross-device compatibility spanning laptops, desktops, IoT, mobile, and cloud, Gemma ensures broadly accessible AI capabilities.

Partnerships with NVIDIA have optimized Gemma for NVIDIA GPUs across various environments, while Google Cloud’s Vertex AI provides a comprehensive MLOps toolset for seamless deployment and customization.

Empowering Researchers and Developers

Google’s commitment to the open AI community is exemplified by its provision of free access to Gemma through Kaggle, Colab notebooks, and Google Cloud credits for researchers and first-time users.

The Unsloth team has further enhanced Gemma 2 2B’s accessibility by releasing quantized 4-bit versions that enable 4x faster downloads for fine-tuning. They have also made fine-tuning 2x faster while using 60% less VRAM, with added support for Flash Attention mechanisms. Unsloth’s efforts extend to creating a user-friendly Chat UI for Gemma-2 Instruct, making the model even more accessible to developers and researchers.

The Path Forward: Open Source AI

As the AI landscape evolves, industry leaders are increasingly recognizing the importance of open source models. Meta CEO Mark Zuckerberg recently outlined why he believes open source AI is the way forward, emphasizing its benefits for developers, organizations, and the world at large.

Zuckerberg argues that open source AI empowers developers to train, fine-tune, and distill models to meet specific needs, ensures control and flexibility, protects sensitive data, and promotes long-term ecosystem investment. Moreover, he asserts that open source AI will be safer than closed alternatives, as transparency enables broader scrutiny and security.

Google’s Gemma 2 vs. Meta’s Llama 3.1

While Meta has made significant strides with the release of Llama 3.1, boasting a 50% cost reduction for inference compared to closed models like GPT-4o, Google’s Gemma 2 series remains a formidable contender.

The Gemma 2 2B model, in particular, has garnered praise for its impressive performance on the LYMSYS arena, surpassing GPT 3.5 and Mixtral 8x7B. Its strong showing on the HF v2 leaderboard, despite struggling in MATH, and its ability to outperform the previous Gemma v1 2B by over 10% in benchmarks highlight the rapid advancements in Google’s open source offerings.

Conclusion

As Google continues to push the boundaries of open source AI with its Gemma 2 series, the tech community eagerly anticipates the innovations and collaborations these releases will inspire. With a steadfast commitment to responsible AI development, compatibility across diverse frameworks and hardware, and empowerment of researchers and developers, Google is well-positioned to lead the charge in the open source AI revolution.

The rapid expansion of the Gemma 2 series, outpacing Meta’s releases, underscores Google’s dedication to democratizing AI and fostering a vibrant open source ecosystem. As the AI landscape evolves, it is clear that open source models will play an increasingly crucial role in shaping the future of this transformative technology.

Categories: AI Tools Guide
X