--- base_model: - mlabonne/gemma-3-12b-it-abliterated - google/gemma-3-12b-it - qiuxi337/gemma-3-12b-it-grpo library_name: transformers tags: - mergekit - merge --- # temp This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the NuSLERP merge method using [mlabonne/gemma-3-12b-it-abliterated](https://huggingface.co/mlabonne/gemma-3-12b-it-abliterated) as a base. Well, it's been a while, hasn't it? It's only been a short 5 months since I posted the last Nemesia version, saying the famous last words > I will try swapping out a model or two in the merge and trying again to upload as a v2.0. Then, exams and university stuff crushed me whole which wasn't great, and I didn't have any time or computational resources to get back to merging on a reasonable timeframe until now. Honestly, in the time since, Qwen2.5-7B really has not turned out to be all that interesting, Mistral Nemo stealing most of its thunder. So, I decided to update my Nemesia mergeset with a new base: **Gemma3-12B**! In my experience, it has been alright, so I'm excited to tinker with it. In my testing, this thing is alright- I couldn't run any GGUFs of it since I don't even know what on earth has happened to my `.venv` in the 5 months past, I'm pretty sure my Transformers installation is corrupted, and I'm re-cloning llama.cpp as I write this. I ran the FP16 and it seemed coherent enough, but I leave that to you to decide. Uses all of the special NuSLERP options because they're there, and is based on the abliterated version of Gemma3 instead of the base because I <3 mlabonne. ### Models Merged The following models were included in the merge: * [google/gemma-3-12b-it](https://huggingface.co/google/gemma-3-12b-it) * [qiuxi337/gemma-3-12b-it-grpo](https://huggingface.co/qiuxi337/gemma-3-12b-it-grpo) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: mlabonne/gemma-3-12b-it-abliterated parameters: weight: 1.0 - model: qiuxi337/gemma-3-12b-it-grpo parameters: weight: 0.5 - model: google/gemma-3-12b-it parameters: weight: 0.2 merge_method: nuslerp base_model: mlabonne/gemma-3-12b-it-abliterated parameters: normalize: true int8_mask: true nuslerp_flatten: false nuslerp_row_wise: true dtype: float16 ```