Models

Gemma 4

Google DeepMind's open model family for local, multimodal, and agentic AI workflows.

Gemma 4 is a family of Apache 2.0 open models from Google DeepMind, designed for reasoning, multimodal inputs, edge deployments, and developer workflows that need more control than hosted-only APIs.

Open sourceApache-2.0Local first
Overview

Gemma 4: what to know first

Gemma 4 is Google DeepMind's latest open model family for developers who want more deployment control than a purely hosted model API can provide. It sits in the growing middle ground between frontier proprietary models and smaller local models: capable enough to test real product workflows, but available in forms that can be evaluated outside a closed chat product.

Gemma 4 is a set of open models released by Google DeepMind under the Apache 2.0 license. The family is positioned for reasoning, multimodal inputs, local and edge use, and agentic workflows. Instead of being a single chatbot, it is a model family that developers can evaluate through official Google surfaces, Hugging Face, and local inference ecosystems.

Open model adoption depends on two things: strong base capability and practical deployment paths. Gemma 4 is important because it pushes both at once. For builders, that means the same project can start with a hosted experiment, move into local testing, and later compare edge or self-hosted deployment without changing the overall model family.

Use cases

Common ways to use it

Local AI product experiments

Use Gemma 4 to test whether a feature can run on local hardware or self-hosted infrastructure before committing to a hosted-only architecture.

Multimodal document and image workflows

The family is relevant for apps that combine text with image understanding, such as visual review, document triage, and assistant-style analysis.

Agentic tool workflows

Gemma 4 is worth evaluating when a workflow needs reasoning across steps, structured outputs, or tool-oriented behavior.

Edge and mobile AI prototypes

The smaller variants are aimed at low-latency and edge use cases where sending every request to a remote model is not ideal.

Fit guide

When it makes sense

Good fit if

  • Developers evaluating open models for local or self-hosted AI apps
  • Teams that need multimodal reasoning without depending only on hosted APIs
  • Builders comparing small edge models against larger workstation-grade open models
  • Researchers and product teams that want Apache 2.0 model weights from a major lab

Not a fit if

  • Users who want a fully managed consumer assistant
  • Teams that do not want to handle model serving, safety testing, or deployment details
  • Workflows that require guaranteed hosted SLA support from the model provider

How to choose

Choose Gemma 4 when open weights and local control matter vs Gemini: Gemini remains Google's flagship hosted model stack, while Gemma 4 is the better fit when you want downloadable weights, local experimentation, and more deployment control.

Compare carefully against other open model families vs Llama, Qwen, Mistral: Gemma 4's appeal is its Google research lineage, Apache 2.0 license, and multimodal direction, but teams should still benchmark it against nearby open models on their own tasks.

Next step

Where to go from here

Try it locally

Command line

Run a workstation model with Ollama

Use this after installing Ollama and confirming your machine has enough memory for the 26B variant.

ollama run gemma4:26b

Run the smaller edge-oriented variant

Use the smaller variant when you want a lighter local test before trying larger Gemma 4 models.

ollama run gemma4:e2b
Technical details Structured profile Facts and source data

At a glance

Status
published
Category
Models
Type
model
License
Apache-2.0
Verified
2026-04-18

Signals

Open sourceLocal firstSelf-hostableAPI

Tags

Category

modelopen source

Capability

local inference

Constraint

open sourceself hostedlocal firstopen weights

Scenario

local aiself hosted ai

Structured outputs

# Gemma 4

Google DeepMind's open model family for local, multimodal, and agentic AI workflows.

## Summary
Gemma 4 is a family of Apache 2.0 open models from Google DeepMind, designed for reasoning, multimodal inputs, edge deployments, and developer workflows that need more control than hosted-only APIs.


## Guide
Gemma 4 is Google DeepMind's latest open model family for developers who want more deployment control than a purely hosted model API can provide. It sits in the growing middle ground between frontier proprietary models and smaller local models: capable enough to test real product workflows, but available in forms that can be evaluated outside a closed chat product.

### What it is
Gemma 4 is a set of open models released by Google DeepMin

Source links

Related

More models