Gemma 4 27B

Googleopen-sourceOpen Source

Google's open-weight flagship under Apache 2.0 — April 2026. Designed for self-host with strong instruction-following and tool calling.

Context Window

128K

tokens

Max Output

8K

tokens

Input Price

Self-host

open weights

Output Price

Self-host

infra cost only

Speed

142

tokens/sec

Released

Apr 2026

2026-04-09

Blended Cost

self-host

Value Score

n/a self-host

Capabilities

ChatFunction CallingCode Generation

Benchmarks

Quality Index
75
MMLU Pro
79.2
HumanEval (Coding)
81.4
MATH
73.6
Arena ELO
1351

Open Source — Licensed under Apache 2.0

About Gemma 4 27B

Gemma 4 27B is a open-source AI model by Google, released on April 9, 2026. It supports a context window of 128K tokens and can generate up to 8K output tokens.

Gemma 4 27B is published as open weights (Apache 2.0) for self-hosting — there is no per-token API price. Cost depends on your inference infrastructure: GPU rental, throughput per GPU, and operating overhead. For commodity 24-48GB GPUs the effective cost typically lands in the $0.10-$0.50 per million output tokens range, well below the cheapest hosted alternatives.

Gemma 4 27B is available as an open-source model under the Apache 2.0 license, meaning you can self-host it for predictable costs or use it through API providers like Swfte Connect.

Using Gemma 4 27B with Swfte

Access Gemma 4 27B through Swfte Connect, our unified LLM gateway. Connect gives you a single API for 50+ models, with automatic routing, cost optimization, and fallback handling. You can also try Gemma 4 27B in our AI Playground before integrating.