--- base_model: [] library_name: transformers tags: - mergekit - merge --- # Nemo-Grand-Gutenburg-bf16-dr This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using G:/11B/Mistral-Nemo-Instruct-2407-12B as a base. ### Models Merged The following models were included in the merge: * G:/11B/Lyra-Gutenberg-mistral-nemo-12B * G:/11B/mistral-nemo-gutenberg-12B-v4 * G:/11B/Lyra4-Gutenberg-12B ### Configuration The following YAML configuration was used to produce this model: ```yaml # # Grand Gutenburg "one step DARK" weights dense. # models: - model: G:/11B/mistral-nemo-gutenberg-12B-v4 parameters: weight: [1,1,1,1,1,1,1,1,1,1,0.7,0.3,0,0,0,0,0,0,0.3,0.7,1,1,1,1,0.7,0.3,0,0,0,0,0,0,0,0,0,0,0,0,0,0] - model: G:/11B/Lyra4-Gutenberg-12B parameters: weight: [0,0,0,0,0,0,0,0,0,0,0.3,0.7,1,1,0.5,0.5,0.5,0.5,0,0,0,0,0,0,0.7,0.3,1,1,1,1,0.7,0.3,0,0,0,0,0,0,0,0] - model: G:/11B/Lyra-Gutenberg-mistral-nemo-12B parameters: weight: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0.5,0.5,0.5,0.5,0.7,0.3,0,0,0,0,0,0,0,0,0,0,0.3,0.7,1,1,1,1,1,1,1,1] merge_method: dare_ties base_model: G:/11B/Mistral-Nemo-Instruct-2407-12B dtype: bfloat16 ```