Quantizations of https://huggingface.co/burtenshaw/GemmaCoder3-12B

Open source inference clients/UIs

Closed source inference clients/UIs


From original readme

This model is a fine-tuned version of google/gemma-3-12b-it on the open-r1/codeforces-cots dataset. It has been trained using TRL.

Evaluation

The model improves performance on the LiveCodeBench benchmark, but diminishes on other benchmarks.

Benchmark GemmaCoder-12B Gemma3-12B-it
Winogrande 63.9% 63.5%
MMLU 61.0% 69.5%
HellaSwag 54.0% 53.5%
LiveCodeBench 32.9% 21.9%

Quick start

from transformers import pipeline

question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
generator = pipeline("text-generation", model="burtenshaw/gemma-3-12b-it-codeforces-SFT", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])

Training procedure

Visualize in Weights & Biases

This model was trained with SFT.

Framework versions

  • TRL: 0.17.0.dev0
  • Transformers: 4.51.0.dev0
  • Pytorch: 2.6.0
  • Datasets: 3.4.1
  • Tokenizers: 0.21.1
Downloads last month
452
GGUF
Model size
11.8B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

1-bit

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support