--- base_model: - djuna/Q2.5-Veltha-14B - Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4 - tanliboy/lambda-qwen2.5-14b-dpo-test - suayptalha/Lamarckvergence-14B - sthenno/tempesthenno-ppo-ckpt40 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE TIES](https://arxiv.org/abs/2311.03099) merge method using [suayptalha/Lamarckvergence-14B](https://huggingface.co/suayptalha/Lamarckvergence-14B) as a base. ### Models Merged The following models were included in the merge: * [djuna/Q2.5-Veltha-14B](https://huggingface.co/djuna/Q2.5-Veltha-14B) * [Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4](https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4) * [tanliboy/lambda-qwen2.5-14b-dpo-test](https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test) * [sthenno/tempesthenno-ppo-ckpt40](https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40) ### Configuration The following YAML configuration was used to produce this model: ```yaml name: OptimalMerge-14B merge_method: dare_ties base_model: suayptalha/Lamarckvergence-14B tokenizer_source: base dtype: bfloat16 out_dtype: bfloat16 parameters: normalize: true int8_mask: true models: - model: suayptalha/Lamarckvergence-14B parameters: weight: 0.35 density: 0.75 - model: sthenno/tempesthenno-ppo-ckpt40 parameters: weight: 0.25 density: 0.7 - model: tanliboy/lambda-qwen2.5-14b-dpo-test parameters: weight: 0.2 density: 0.65 - model: djuna/Q2.5-Veltha-14B parameters: weight: 0.1 density: 0.6 - model: Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4 parameters: weight: 0.1 density: 0.6 adaptive_merge_parameters: task_weights: IFEval: 1.8 BBH: 1.5 MATH: 2.0 GPQA: 1.4 MUSR: 1.3 MMLU-PRO: 1.5 smoothing_factor: 0.1 gradient_clipping: suayptalha/Lamarckvergence-14B: 0.85 sthenno/tempesthenno-ppo-ckpt40: 0.88 tanliboy/lambda-qwen2.5-14b-dpo-test: 0.87 djuna/Q2.5-Veltha-14B: 0.89 Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4: 0.86 ```