--- base_model: - Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1 - nvidia/AceMath-1.5B-Instruct - Qwen/Qwen2.5-1.5B-Instruct - bond005/meno-tiny-0.1 - hotmailuser/QwenStock-1.7B - fblgit/miniclaus-qw1.5B-UNAMGS-GRPO - Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2 - Sakalti/Saba1-1.8B - hotmailuser/QwenModelStock-1.8B - Sakalti/SJT-1.5B-Alpha - lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3 - nvidia/AceInstruct-1.5B - Sakalti/Saba1.5-1.5B - fblgit/miniclaus-qw1.5B-UNAMGS library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1](https://huggingface.co/Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1) * [nvidia/AceMath-1.5B-Instruct](https://huggingface.co/nvidia/AceMath-1.5B-Instruct) * [bond005/meno-tiny-0.1](https://huggingface.co/bond005/meno-tiny-0.1) * [hotmailuser/QwenStock-1.7B](https://huggingface.co/hotmailuser/QwenStock-1.7B) * [fblgit/miniclaus-qw1.5B-UNAMGS-GRPO](https://huggingface.co/fblgit/miniclaus-qw1.5B-UNAMGS-GRPO) * [Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2](https://huggingface.co/Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2) * [Sakalti/Saba1-1.8B](https://huggingface.co/Sakalti/Saba1-1.8B) * [hotmailuser/QwenModelStock-1.8B](https://huggingface.co/hotmailuser/QwenModelStock-1.8B) * [Sakalti/SJT-1.5B-Alpha](https://huggingface.co/Sakalti/SJT-1.5B-Alpha) * [lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3](https://huggingface.co/lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3) * [nvidia/AceInstruct-1.5B](https://huggingface.co/nvidia/AceInstruct-1.5B) * [Sakalti/Saba1.5-1.5B](https://huggingface.co/Sakalti/Saba1.5-1.5B) * [fblgit/miniclaus-qw1.5B-UNAMGS](https://huggingface.co/fblgit/miniclaus-qw1.5B-UNAMGS) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: bond005/meno-tiny-0.1 - model: fblgit/miniclaus-qw1.5B-UNAMGS - model: fblgit/miniclaus-qw1.5B-UNAMGS-GRPO - model: hotmailuser/QwenModelStock-1.8B - model: hotmailuser/QwenStock-1.7B - model: lalainy/ECE-PRYMMAL-YL-1B-SLERP-V3 - model: nvidia/AceInstruct-1.5B - model: nvidia/AceMath-1.5B-Instruct - model: Qwen/Qwen2.5-1.5B-Instruct - model: Sakalti/Saba1.5-1.5B - model: Sakalti/Saba1-1.8B - model: Sakalti/SJT-1.5B-Alpha - model: Youlln/ECE-PRYMMAL-YL-1B-SLERP-V1 - model: Youlln/ECE-PRYMMAL-YL-1B-SLERP-V2 tokenizer: source: base merge_method: sce base_model: Qwen/Qwen2.5-1.5B-Instruct dtype: bfloat16 parameters: int8_mask: true ```