tssst's picture
Upload 17 files
6f32206 verified
---
base_model:
- mlabonne/gemma-3-12b-it-abliterated
- google/gemma-3-12b-it
- qiuxi337/gemma-3-12b-it-grpo
library_name: transformers
tags:
- mergekit
- merge
---
# temp
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the NuSLERP merge method using [mlabonne/gemma-3-12b-it-abliterated](https://huggingface.co/mlabonne/gemma-3-12b-it-abliterated) as a base.
Well, it's been a while, hasn't it? It's only been a short 5 months since I posted the last Nemesia version, saying the famous last words
> I will try swapping out a model or two in the merge and trying again to upload as a v2.0.
Then, exams and university stuff crushed me whole which wasn't great, and I didn't have any time or computational resources to get back to merging on a reasonable timeframe until now.
Honestly, in the time since, Qwen2.5-7B really has not turned out to be all that interesting, Mistral Nemo stealing most of its thunder. So, I decided to update my Nemesia mergeset with a
new base: **Gemma3-12B**! In my experience, it has been alright, so I'm excited to tinker with it.
In my testing, this thing is alright- I couldn't run any GGUFs of it since I don't even know what on earth has happened to my `.venv` in the 5 months past, I'm pretty sure my Transformers installation
is corrupted, and I'm re-cloning llama.cpp as I write this. I ran the FP16 and it seemed coherent enough, but I leave that to you to decide.
Uses all of the special NuSLERP options because they're there, and is based on the abliterated version of Gemma3 instead of the base because I <3 mlabonne.
### Models Merged
The following models were included in the merge:
* [google/gemma-3-12b-it](https://huggingface.co/google/gemma-3-12b-it)
* [qiuxi337/gemma-3-12b-it-grpo](https://huggingface.co/qiuxi337/gemma-3-12b-it-grpo)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: mlabonne/gemma-3-12b-it-abliterated
parameters:
weight: 1.0
- model: qiuxi337/gemma-3-12b-it-grpo
parameters:
weight: 0.5
- model: google/gemma-3-12b-it
parameters:
weight: 0.2
merge_method: nuslerp
base_model: mlabonne/gemma-3-12b-it-abliterated
parameters:
normalize: true
int8_mask: true
nuslerp_flatten: false
nuslerp_row_wise: true
dtype: float16
```