Gemma 4 Model Family Overview
Gemma 4 is Google DeepMind's April 2026 open model family with four variants (E2B, E4B, 26B MoE, 31B Dense), now Apache 2.0 licensed, spanning edge devices to top-tier reasoning.
Gemma 4 is Google DeepMind's latest open model family, released April 2, 2026. It is the first Gemma release under the Apache 2.0 license (previous versions used a custom Google license), allowing commercial use, forking, modification, and redistribution with attribution. The lineup includes four models: - E2B (Effective 2B): ~5GB at Q8, 128K context, text + image + audio + video. Designed for phones and edge devices. - E4B (Effective 4B): ~8GB at Q8, 128K context, text + image + audio + video. Mobile and lightweight compute. - 26B MoE (Mixture of Experts, 4B active params): 256K context, text + image only. Fast inference, ~18GB at 4-bit. - 31B Dense: ~35GB at Q8, 256K context, text + image only. Highest quality, ranked #3 open model globally on Arena AI. All models are thinking models by default. Thinking can be disabled by omitting the <|think|> token in the system prompt.