GemmaDenseGemma License (Apache 2.0 compatible)

Gemma 3 27B

Gemma 3 27B is the largest Gemma 3 model — a dense 27B with vision support, 128K context, and strong multilingual coverage. At Q4_K_M it needs ~15 GB VRAM, fitting on 16 GB GPUs with limited context or 24 GB with room to spare. The GeGLU ac

27.0B

Parameters

128K

Max Context

Dense

Architecture

Mar 12, 2025

Released

Text + Vision

Modality

About Gemma 3 27B

Gemma 3 27B is the largest Gemma 3 model — a dense 27B with vision support, 128K context, and strong multilingual coverage. At Q4_K_M it needs ~15 GB VRAM, fitting on 16 GB GPUs with limited context or 24 GB with room to spare. The GeGLU activation and unique head_dim=176 give it a distinctive performance profile. Strong across general knowledge, coding, and multilingual tasks. A solid alternative to Qwen 2.5 32B for users who prefer Google's training methodology.

General PurposeMultilingualVisionCode

Technical Specifications

Total Parameters27.0B
ArchitectureDense
Attention TypeGQA (Grouped Query Attention)
Hidden Dimensiond = 5,632
Transformer Layers64
Attention Heads32
KV Headsn_kv = 8
Head Dimensiond_head = 176
Activation FunctionGeGLU
NormalizationRMSNorm
Position EmbeddingRoPE

System Requirements

Estimated VRAM at 10% overhead for different quantization methods and context sizes.

Quantization1K ctx128K ctx
Q4_K_M0.50 B/W
~97% of FP16
14.30Consumer GPU
57.96Datacenter GPU
Q8_01.00 B/W
~100% of FP16
28.26Datacenter GPU
71.91Datacenter GPU
F162.00 B/W
Reference
56.17Datacenter GPU
99.82Cluster / Multi-GPU
Fits 24 GB consumer GPU
Fits 80 GB datacenter GPU
Requires cluster / multi-GPU

Other Gemma Models

View All

Find the right GPU for Gemma 3 27B

Use the interactive VRAM Calculator to see exactly how much memory you need at any quantization level, context length, and overhead setting.