merge
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
models:
- model: mlabonne/gemma-3-4b-it-abliterated
- model: mshojaei77/gemma-3-4b-persian-v0
base_model: mlabonne/gemma-3-4b-it-abliterated
merge_method: slerp
dtype: bfloat16 # Better stability for precision-sensitive merges
parameters:
density: 0.5
weight:
- filter: "self_attn"
value: [0.75, 0.4, 0.25, 0.4, 0.75] # U-shaped attention weighting
- filter: "mlp"
value: [0.25, 0.6, 0.9, 0.6, 0.25] # ฮ-shaped MLP weighting
t: [0.15, 0.35, 0.65, 0.35, 0.15] # Optimized linguistic injection
generation_config = {
"temperature": 1.1,
"top_k": 50,
"top_p": 0.9,
"repetition_penalty": 1.15,
"do_sample": True
}
- Downloads last month
- 11
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for mshojaei77/gemma-3-4b-persian-v0-abliterated
Merge model
this model