RUN THIS LLM

Search local LLM hardware requirements

← Back to all models

Gemma 4 27B MoE

Google · 26B MoE · Vision

Gemma 4 sparse MoE with 26B total params and only 4B active per token. Flagship-level multimodal quality at efficient inference cost.

VRAM Requirements

QuantizationVRAM
Q4_K_M (smallest)15.6 GB
Q8_0 (balanced)28.6 GB
FP16 (full quality)52 GB

Specifications

Benchmarks

Loading interactive analysis...

RTL
Add Run This LLM to your home screen