YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

MoE-Girl-1BA-7BT - bnb 8bits

Original model description:

library_name: transformers license: apache-2.0 base_model: allenai/OLMoE-1B-7B-0924 tags: - axolotl - moe - roleplay model-index: - name: MoE-girl_1BA_7BT results: []

MoE Girl 1bA 7bT

image/jpeg A finetune of OLMoE by AllenAI designed for roleplaying (and maybe general usecases if you try hard enough).

Disclaimer

PLEASE do not expect godliness out of this, it's a model with 1 billion active parameters. Expect something more akin to Gemma 2 2B, not Llama 3 8B.

Quants

GGUF (requires a newish version of llama.cpp or kobold.cpp 1.76):

Prompting

Use ChatML.

<|im_start|>system
You are a helpful assistant who talks like a pirate.<|im_end|>
<|im_start|>user
Hello there!<|im_end|>
<|im_start|>assistant
Yarr harr harr, me matey!<|im_end|>

Thanks

Special thanks to the members of Allura for testing and emotional support, as well as the creators of all the datasets that were used in the Special Sauce used to train this model. I love you all <3 - Fizz

Downloads last month
2
Safetensors
Model size
6.92B params
Tensor type
F32
FP16
I8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support