NekoMix-12B-exl2

Original model: NekoMix-12B by Moraliane

Quants

4bpw h6 (main)
4.5bpw h6
5bpw h6
6bpw h6
8bpw h8

Quantization notes

Made with Exllamav2 0.2.8 with default dataset.
It seems to be primarily a Russian RP model. No clue how it performs at all.
It can be used with TabbyAPI or Text-Generation-WebUI with RTX GPU on Windows or RTX/ROCm on Linux.
Exllamav2 doesn't support offloading to RAM, so make sure it fits your GPU. Otherwise use GGUF quants instead.
For example, with 12GB VRAM it can be used at 6bpw/Q6 cache at 16k context.

Эта модель может использоваться с TabbyAPI или Text-Generation-WebUI.
Для работы с ней требуется Nvidia RTX (Windows) или RTX/ROCm (Linux).
Exl2 формат требует, чтобы модель полностью помещалась в видеопамяти.
Например, с 12ГБ видеопамяти можно использовать 6bpw версию с Q6 кэшем с 16k контекстом.

Original model card

NekoMix-12B

NekoMix-12B

GGUF:

https://huggingface.co/mradermacher/NekoMix-12B-GGUF

GGUF imatrix:

Soon...

Presets:

https://huggingface.co/Moraliane/NekoMix-12B/blob/main/pres/NekoMixRUS.json NekoMix-12B NekoMix-12B Так же рекомендую использовать Mistral V3-Tekken в качестве Context Template и Instruct Template (!!Спорно!!)

Sampler:

NekoMix-12B

NekoMix-12B

Рекомендую для начала использовать стоковый пресет simple-1 а так же Parameters_Top(A)Kek из https://huggingface.co/MarinaraSpaghetti/SillyTavern-Settings/tree/main/Parameters

Temp - 0,7 - 1,25 ~
TopA - 0,1
DRY - 0,8 1,75 2 0
I recommend trying the stock presets from SillyTavern, such as simple-1.

Testmrg

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the della_linear merge method using E:\Programs\TextGen\text-generation-webui\models\IlyaGusev_saiga_nemo_12b as a base.

Models Merged

The following models were included in the merge:

  • E:\Programs\TextGen\text-generation-webui\models\MarinaraSpaghetti_NemoMix-Unleashed-12B
  • E:\Programs\TextGen\text-generation-webui\models\Vikhrmodels_Vikhr-Nemo-12B-Instruct-R-21-09-24
  • E:\Programs\TextGen\text-generation-webui\models\TheDrummer_Rocinante-12B-v1.1

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: E:\Programs\TextGen\text-generation-webui\models\IlyaGusev_saiga_nemo_12b
    parameters:
      weight: 0.5  # Основной акцент на русском языке
      density: 0.4
  - model: E:\Programs\TextGen\text-generation-webui\models\MarinaraSpaghetti_NemoMix-Unleashed-12B
    parameters:
      weight: 0.2  # РП модель, чуть меньший вес из-за ориентации на английский
      density: 0.4
  - model: E:\Programs\TextGen\text-generation-webui\models\TheDrummer_Rocinante-12B-v1.1
    parameters:
      weight: 0.2  # Увеличенный вес для усиления РП аспектов
      density: 0.5  # Повышенная плотность для более сильного влияния
  - model: E:\Programs\TextGen\text-generation-webui\models\Vikhrmodels_Vikhr-Nemo-12B-Instruct-R-21-09-24
    parameters:
      weight: 0.25  # Русскоязычная поддержка и баланс
      density: 0.4

merge_method: della_linear
base_model: E:\Programs\TextGen\text-generation-webui\models\IlyaGusev_saiga_nemo_12b
parameters:
  epsilon: 0.05
  lambda: 1
dtype: float16
tokenizer_source: base
Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for cgus/NekoMix-12B-exl2

Quantized
(5)
this model