Output
This is a merge of pre-trained language models created using mergekit.
RP/PROSE MODEL MERGE BASED ON NEMO-MISTRAL
Chat Template: ChatML (or V3 TEKKEN)
GGUF Quants: https://huggingface.co/mradermacher/Kaolinite-Kitara-12B-GGUF
Imatrix Quants: https://huggingface.co/mradermacher/Kaolinite-Kitara-12B-i1-GGUF
Merge Details
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
- Chimera-Lotus-12B
- Kaolinite-12B
Configuration
The following YAML configuration was used to produce this model:
merge_method: slerp
base_model: Kaolinite-12B
models:
- model: Kaolinite-12B
- model: Chimera-Lotus-12B
parameters:
t: [0.8, 0.8, 0.7, 0.7, 0.6, 0.6, 0.5, 0.5, 0.4, 0.4,
0.7, 0.6, 0.6, 0.5, 0.5, 0.4, 0.4, 0.3, 0.3, 0.2,
0.5, 0.4, 0.4, 0.3, 0.3, 0.2, 0.2, 0.1, 0.1, 0.0,
0.3, 0.2, 0.2, 0.1, 0.1, 0.0, 0.0, 0.0, 0.0, 0.0]
dtype: bfloat16
normalize: true
tokenizer:
source: "union"
tokens: {}
chat_template: "chatml"
output_model_name: "Kaolinite-Kitara-12B"
- Downloads last month
- 20
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support