Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

RareBit-v2-32B

Another big merge, similar in idea to RP-Stew. V2 here hasn't dropped a random Chinese character like V1 did yet after 100 swipes, which might be because I regulated QwQ to only being used as the base model, instead of mixing it wholesale. Only other change was using v4 of ArliAI's model in the mix. I still need to do some more testing with it to see if it's fully ready to be shared in a broader sense, but so far it's been pretty good. I'll make a proper model page later next week, but this is what I've gathered from it so far:

Pros:

  • Prose seem natural and creative.
  • Hasn't made any big logical mistakes.
  • Stays in-character and hasn't responded as user.
  • Decent thinking capabilities.
  • No refusals, even during the thinking stage.

Cons:

  • None so far from testing, but I doubt it's perfect. I'm sure there's something I missed, so consider this pending full critique.

Big thanks to the original model creators for providing the ingredients!

  • Qwen
  • EVA-UNIT-01
  • arcee-ai
  • ArliAI
  • trashpanda

GGUF (provided by mradermacher)

https://huggingface.co/mradermacher/RareBit-v2-32B-GGUF

https://huggingface.co/mradermacher/RareBit-v2-32B-i1-GGUF

Prompt Format: ChatML

<|im_start|>system
System prompt<|im_end|>
<|im_start|>user
User prompt<|im_end|>
<|im_start|>assistant
Bot response<|im_end|>

Models Merged

The following models were included in the merge:

https://huggingface.co/Qwen/QwQ-32B

https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2

https://huggingface.co/arcee-ai/Virtuoso-Medium-v2

https://huggingface.co/ArliAI/QwQ-32B-ArliAI-RpR-v4

https://huggingface.co/trashpanda-org/QwQ-32B-Snowdrop-v0

Downloads last month
21
Safetensors
Model size
9.85B params
Tensor type
F16
·
I16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for async0x42/RareBit-v2-32B-exl3_4.5bpw