This model was quantized using commit 3a90264 of the dev branch of exllamav2. The 8k context bug looks to be thoroughly squashed as of this commit. To use this model, please either build your own copy of exllamav2 from the dev branch, or wait for the forthcoming v0.2.9 release.
The original model can be found here.
✨G3 Glitter 27B✨

A creative writing model based on Gemma 3 27B.
Columbidae/gemma-3-27b-half, a 50/50 merge of 27B IT and 27B PT, was used as the base model. (This was done because of the success of Starshine, a 50/50 IT and PT merge.)
The inclusion of PT model does weaken the instruct, but it also weakens the censorship/hesitancy to participate in certain fictional stories. The prose also becomes more natural with less of the IT model included.
This model does better with short and to-the-point prompts. Long, detailed system prompts will often confuse it. (Tested with 1000-2000 token system prompts to lackluster results compared to 100-500 token prompts).
Instruct Format
Uses Gemma2/3 instruct and context. Like Glitter 12b, this works well with temp = 1, top-nsigma = 1.5
.
<start_of_turn>user
{User messages; can also put sysprompt here to use the built-in g3 training}<end_of_turn>
<start_of_turn>model
{model response}<end_of_turn>
- Downloads last month
- 2
Model tree for MikeRoz/allura-org_Gemma-3-Glitter-27B-4.0bpw-h6-exl2
Base model
allura-org/Gemma-3-Glitter-27B