--- base_model: - THUDM/LongReward-glm4-9b-DPO - allura-org/GLM4-9B-Neon-v2 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [allura-org/GLM4-9B-Neon-v2](https://huggingface.co/allura-org/GLM4-9B-Neon-v2) as a base. ### Models Merged The following models were included in the merge: * [THUDM/LongReward-glm4-9b-DPO](https://huggingface.co/THUDM/LongReward-glm4-9b-DPO) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: allura-org/GLM4-9B-Neon-v2 - model: THUDM/LongReward-glm4-9b-DPO parameters: weight: [0.496, 0.166, 0.166, 0.496, 0.496, 0.166, 0.166, 0.496] base_model: allura-org/GLM4-9B-Neon-v2 merge_method: sce parameters: select_topk: 0.06 lambda: 0.66 tokenizer_source: base dtype: float32 out_dtype: bfloat16 ```