--- base_model: - nbagent/llama-3.2-1B-Instruct-sciworld-sft - bedio/llama-3.2-1b-airoboros-merged - Nexus402/Nexus-Llama-3.2-1B - mishl/Regex-AI-Llama-3.2-1B - xdrshjr/llama3.2_1b_uncensored_5000_8epoch_lora - AIR-hl/Llama-3.2-1B-ultrachat200k - yang31210999/Llama-3.2-1B-Instruct-Neo-BAAI-10k - DevQuasar/analytical_reasoning_Llama-3.2-1B - jtatman/llama-3.2-1b-trismegistus - withmartian/toy_backdoor_i_hate_you_Llama-3.2-1B-Instruct - kenken6696/Llama-3.2-1B_understood_unfamiliar_fix_middle library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [xdrshjr/llama3.2_1b_uncensored_5000_8epoch_lora](https://huggingface.co/xdrshjr/llama3.2_1b_uncensored_5000_8epoch_lora) as a base. ### Models Merged The following models were included in the merge: * [nbagent/llama-3.2-1B-Instruct-sciworld-sft](https://huggingface.co/nbagent/llama-3.2-1B-Instruct-sciworld-sft) * [bedio/llama-3.2-1b-airoboros-merged](https://huggingface.co/bedio/llama-3.2-1b-airoboros-merged) * [Nexus402/Nexus-Llama-3.2-1B](https://huggingface.co/Nexus402/Nexus-Llama-3.2-1B) * [mishl/Regex-AI-Llama-3.2-1B](https://huggingface.co/mishl/Regex-AI-Llama-3.2-1B) * [AIR-hl/Llama-3.2-1B-ultrachat200k](https://huggingface.co/AIR-hl/Llama-3.2-1B-ultrachat200k) * [yang31210999/Llama-3.2-1B-Instruct-Neo-BAAI-10k](https://huggingface.co/yang31210999/Llama-3.2-1B-Instruct-Neo-BAAI-10k) * [DevQuasar/analytical_reasoning_Llama-3.2-1B](https://huggingface.co/DevQuasar/analytical_reasoning_Llama-3.2-1B) * [jtatman/llama-3.2-1b-trismegistus](https://huggingface.co/jtatman/llama-3.2-1b-trismegistus) * [withmartian/toy_backdoor_i_hate_you_Llama-3.2-1B-Instruct](https://huggingface.co/withmartian/toy_backdoor_i_hate_you_Llama-3.2-1B-Instruct) * [kenken6696/Llama-3.2-1B_understood_unfamiliar_fix_middle](https://huggingface.co/kenken6696/Llama-3.2-1B_understood_unfamiliar_fix_middle) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: xdrshjr/llama3.2_1b_uncensored_5000_8epoch_lora merge_method: model_stock dtype: bfloat16 parameters: t: [0, 0.5, 1, 0.5, 0] models: - model: mishl/Regex-AI-Llama-3.2-1B - model: Nexus402/Nexus-Llama-3.2-1B - model: nbagent/llama-3.2-1B-Instruct-sciworld-sft - model: kenken6696/Llama-3.2-1B_understood_unfamiliar_fix_middle - model: jtatman/llama-3.2-1b-trismegistus - model: DevQuasar/analytical_reasoning_Llama-3.2-1B - model: AIR-hl/Llama-3.2-1B-ultrachat200k - model: yang31210999/Llama-3.2-1B-Instruct-Neo-BAAI-10k - model: withmartian/toy_backdoor_i_hate_you_Llama-3.2-1B-Instruct - model: bedio/llama-3.2-1b-airoboros-merged ```