Google Gemma 4 Officially Released: Four Sizes Under Apache 2.0 License, 31B Global Open Source Third

BlockBeatNews

According to 1M AI News monitoring, Google has officially released the open-source model family Gemma 4, which includes 4 models of different sizes, all under the Apache 2.0 license. Google says this is a response to feedback from the community. Apache 2.0 means developers are free to use commercially, modify, and distribute without additional restrictions. Hugging Face co-founder and CEO Clément Delangue calls this a “huge milestone.”

The 4 models are designed for different hardware scenarios:

  1. 31B Dense: the highest raw quality, suitable for fine-tuning; unquantized weights can run on a single 80GB H100, while the quantized version supports consumer-grade GPUs
  2. 26B MoE (Mixture of Experts): a total of 26B parameters, but only 3.8B are activated during inference; it focuses on low latency and is suitable for agent scenarios where speed matters
  3. E4B and E2B: edge models for mobile phones and IoT devices; about 4B and 2B parameters are activated during inference, respectively—enough to maintain device battery life; native support for audio input (speech recognition and understanding) and can run fully offline

On the text leaderboard of Arena AI, a large-model anonymous battle evaluation platform, the 31B ranks third among global open-source models, while the 26B ranks sixth. Google says it is a “model that surpasses those 20 times its size.” The models are built based on the same research and technology as Gemini 3.

Core capabilities include multi-step reasoning and planning, native function calling and JSON structured output (for agent workflows), code generation, and image and video understanding (across the whole series), along with native training in over 140 languages. Edge models support a 128K context window, while the large models support up to 256K. E2B and E4B, co-optimized with Google Pixel teams, Qualcomm, and MediaTek, can run on devices such as phones, Raspberry Pi, and NVIDIA Jetson Orin Nano. Android developers can build agent applications via the AICore Developer Preview prototype, preparing for compatibility with the future Gemini Nano 4.

In terms of the ecosystem, on day one it already supports major frameworks such as Hugging Face, vLLM, llama.cpp, MLX, Ollama, NVIDIA NIM, LM Studio, and Unsloth. It can be experienced directly in Google AI Studio (31B and 26B) and AI Edge Gallery (E4B and E2B). Since the first release, the Gemma series has been downloaded more than 400 million times, with over 100,000 community-derived variants.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.
Comment
0/400
No comments