Learn how to run & fine-tune Gemma 3n correctly - Read our Guide.

See our collection for all versions of Gemma 3n including GGUF, 4-bit & 16-bit formats.

Unsloth Dynamic 2.0 achieves SOTA accuracy & performance versus other quants.

✨ Gemma 3n Usage Guidelines

  • Currently only text is supported.
  • Ollama: ollama run hf.co/unsloth/gemma-3n-E4B-it:Q4_K_XL - auto-sets correct chat template and settings
  • Set temperature = 1.0, top_k = 64, top_p = 0.95, min_p = 0.0
  • Gemma 3n max tokens (context length): 32K. Gemma 3n chat template:
<bos><start_of_turn>user\nHello!<end_of_turn>\n<start_of_turn>model\nHey there!<end_of_turn>\n<start_of_turn>user\nWhat is 1+1?<end_of_turn>\n<start_of_turn>model\n

🦥 Fine-tune Gemma 3n with Unsloth

Unsloth supports Free Notebooks Performance Memory use
Gemma-3n-E4B ▶️ Start on Colab 2x faster 80% less
GRPO with Gemma 3 (1B) ▶️ Start on Colab 2x faster 80% less
Gemma 3 (4B) ▶️ Start on Colab 2x faster 60% less
Qwen3 (14B) ▶️ Start on Colab 2x faster 60% less
DeepSeek-R1-0528-Qwen3-8B (14B) ▶️ Start on Colab 2x faster 80% less
Llama-3.2 (3B) ▶️ Start on Colab 2.4x faster 58% less

Gemma-3n-E4B model card

Downloads last month
90
Safetensors
Model size
8.39B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for unsloth/gemma-3n-E4B

Finetuned
(1)
this model

Collection including unsloth/gemma-3n-E4B