name: OptimalMerge-14B | |
merge_method: dare_ties | |
base_model: suayptalha/Lamarckvergence-14B | |
tokenizer_source: base | |
dtype: bfloat16 | |
out_dtype: bfloat16 | |
parameters: | |
normalize: true | |
int8_mask: true | |
models: | |
- model: suayptalha/Lamarckvergence-14B | |
parameters: | |
weight: 0.35 | |
density: 0.75 | |
- model: sthenno/tempesthenno-ppo-ckpt40 | |
parameters: | |
weight: 0.25 | |
density: 0.7 | |
- model: tanliboy/lambda-qwen2.5-14b-dpo-test | |
parameters: | |
weight: 0.2 | |
density: 0.65 | |
- model: djuna/Q2.5-Veltha-14B | |
parameters: | |
weight: 0.1 | |
density: 0.6 | |
- model: Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4 | |
parameters: | |
weight: 0.1 | |
density: 0.6 | |
adaptive_merge_parameters: | |
task_weights: | |
IFEval: 1.8 | |
BBH: 1.5 | |
MATH: 2.0 | |
GPQA: 1.4 | |
MUSR: 1.3 | |
MMLU-PRO: 1.5 | |
smoothing_factor: 0.1 | |
gradient_clipping: | |
suayptalha/Lamarckvergence-14B: 0.85 | |
sthenno/tempesthenno-ppo-ckpt40: 0.88 | |
tanliboy/lambda-qwen2.5-14b-dpo-test: 0.87 | |
djuna/Q2.5-Veltha-14B: 0.89 | |
Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4: 0.86 |