--- base_model: - Qwen/Qwen2.5-0.5B-Instruct - Qwen/Qwen2-0.5B-Instruct library_name: transformers tags: - mergekit - merge --- # What? This model is surprisingly usable, despite being an abomination. # How? Mergekit GUI, although it is my own duplicate, uploading to Stupidity-AI instead. # Why? No clue. # Who? Me. # Where? USA. # When? Check the "last updated" field, stupid. # ***ORIGINAL README:*** # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Task Arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [Qwen/Qwen2-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: Qwen/Qwen2.5-0.5B-Instruct dtype: bfloat16 merge_method: task_arithmetic parameters: normalize: true slices: - sources: - layer_range: [0, 24] model: Qwen/Qwen2.5-0.5B-Instruct - layer_range: [0, 24] model: Qwen/Qwen2.5-0.5B-Instruct parameters: weight: [1.0, -1.0, 1.0, -1.0, 1.0] - layer_range: [0, 24] model: Qwen/Qwen2-0.5B-Instruct parameters: weight: [-1.0, 1.0, -1.0, 1.0, -1.0] ```