Google DeepMind releases Gemma 4, its latest open-weights model

Google DeepMind released Gemma 4, a new family of open-weight models under Apache 2.0 license. Available now.

gemma 4 featured

Google DeepMind released Gemma 4 today, a new family of open-weight models built on technology from Gemini 3. The weights are free to download, fine-tune, and deploy commercially under an Apache 2.0 license.

The family comes in four sizes, split across two different use cases.

The larger models (26B and 31B) are built for workstations and consumer GPUs. The 26B uses a Mixture of Experts architecture, meaning only about 4 billion parameters are active during inference, keeping memory requirements lower than a dense model of similar capability. The 31B is a full dense model aimed at tasks that need deeper reasoning.

The edge models (E2B and E4B) target mobile and IoT devices. Both support multimodal inputs — text, image, and audio — and are designed to run completely offline on hardware like phones, Raspberry Pi, and Jetson Nano.

  • Context window: Up to 256K tokens, double what Gemma 3 supported
  • Languages: 140+
  • Key capabilities: Function calling, structured JSON output, agentic workflows, multimodal reasoning

Where to get it:

The instruction-tuned variants of the 26B and 31B models are live now.

RunPod
RunPod

If you need on-demand GPUs for training, fine-tuning, inference, or running open-source models, give RunPod a try.

  • Available hardware: H100, H200, A100, L40S, RTX 4090, RTX 5090, and 30+ more
  • Cost: significantly cheaper than AWS or GCP, billed per second, no contracts
  • Setup: spins up in under a minute, 30+ regions worldwide
Try RunPod →
Affiliate disclosure: We may earn a commission if you sign up via our link, at no extra cost to you.
Efficienist Newsletter

Get the core business tech news delivered straight to your inbox. We track AI, automation, SaaS, and cybersecurity so you don't have to.

Just read what you want, and be done with it.

Read Next