ArtusDev/allura-org_Gemma-3-Glitter-4B-GGUF
โจ Overview :3
This repository contains GGUF format model files converted from allura-org/Gemma-3-Glitter-4B.
The conversion was performed by ArtusDev using llama.cpp
, specifically utilizing the imatrix
quantization option for potentially improved performance.
๐ Original Model Details ^_^
For more information about the model please refer to the original model card. It's pretty neat (empty)!
๐ฌ Instruct Format >.<
This model uses a custom Gemma 2/3 instruct format. It has been trained to recognize an optional system
role.
<start_of_turn>system
{optional system prompt here}<end_of_turn>
<start_of_turn>user
{User messages. You can also place the system prompt here.}<end_of_turn>
<start_of_turn>model
{Model's response}<end_of_turn>
Note: Always ensure the format strictly adheres to the required tokens and structure for optimal model performance. Don't mess it up :3!
GGUF Quantizations (imatrix) by ArtusDev >:3
- Downloads last month
- 975
Hardware compatibility
Log In
to view the estimation
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no library tag.
Model tree for ArtusDev/allura-org_Gemma-3-Glitter-4B-GGUF
Base model
google/gemma-3-4b-pt
Finetuned
google/gemma-3-4b-it
Finetuned
allura-org/Gemma-3-Glitter-4B