Output

This is a merge of pre-trained language models created using mergekit.

RP/PROSE MODEL MERGE BASED ON NEMO-MISTRAL

Chat Template: ChatML (or V3 TEKKEN)

GGUF Quants: https://huggingface.co/mradermacher/Kaolinite-Kitara-12B-GGUF

Imatrix Quants: https://huggingface.co/mradermacher/Kaolinite-Kitara-12B-i1-GGUF

Merge Details

Merge Method

This model was merged using the SLERP merge method.

Models Merged

The following models were included in the merge:

  • Chimera-Lotus-12B
  • Kaolinite-12B

Configuration

The following YAML configuration was used to produce this model:

merge_method: slerp
base_model: Kaolinite-12B
models:
      - model:  Kaolinite-12B
      - model:  Chimera-Lotus-12B
       
parameters:
  t: [0.8, 0.8, 0.7, 0.7, 0.6, 0.6, 0.5, 0.5, 0.4, 0.4,
      0.7, 0.6, 0.6, 0.5, 0.5, 0.4, 0.4, 0.3, 0.3, 0.2,
      0.5, 0.4, 0.4, 0.3, 0.3, 0.2, 0.2, 0.1, 0.1, 0.0,
      0.3, 0.2, 0.2, 0.1, 0.1, 0.0, 0.0, 0.0, 0.0, 0.0]
dtype: bfloat16
normalize: true
tokenizer:
  source: "union"
  tokens: {}
  chat_template: "chatml"
output_model_name: "Kaolinite-Kitara-12B"
Downloads last month
20
Safetensors
Model size
12.2B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for XCryptoniusX/Kaolinite-Kitara-12B

Merges
1 model
Quantizations
2 models