Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

Chuluun-Qwen2.5-72B-v0.08

image/png

This is a merge of pre-trained language models created using mergekit.

I re-ran the original Chuluun formula including the newly released Ink from Allura-Org. I've found the addition gives the model a lot more variability, likely because of aggressive de-slop applied to its dataset. Sometimes this means a word choice will be strange and you'll want to manually edit when needed, but it means you'll see less ministrations sparkling with mischief.

Because of this the best way to approach the model is to run multiple regens and choose the one you like, edit mercilessly, and continue. Like the original Chuluun this variant is very steerable for complex storywriting and RP. It's probably also a little spicier than v0.01 with both Magnum and whatever the heck Fizz threw into the data for Ink.

I've also been hearing praise for a level of character intelligence not seen in other models, including Largestral finetunes and merges. I'm not about to say any model of mine is smarter because it was a dumb idea to use Tess as the base and it somehow worked.

Tips for effective use

As with all writing-focused models balancing intelligence with creativity is tricky. If this one seems like it understands some details but not others try v0.01, overall I think this model is more creative but a little less coherent. If v0.08 is a little too chaotic for your tastes consider running v0.01 at first and switch to this model mid-story if it gets stale.

All the models within the merge use ChatML format so you'll want to use it too. Use Konnect's Qwenception prompt or whatever you prefer, it seems to do fine with any decent sysprompt. Lower temps are suggested than with v0.01 because of Ink's de-slop dataset, testers reported anywhere between 1 and 1.2 as a baseline but plan to adjust to taste. Consider dynatemp for this model as well. If dialogue gets repetitive at all that's usually a sign you need more temp.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using migtissera/Tess-v2.5.2-Qwen2-72B as a base.

Models Merged

The following models were included in the merge:

  • EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2
  • Sao10K/72B-Qwen2.5-Kunou-v1
  • anthracite-org/magnum-v4-72b
  • allura-org/Qwen2.5-72b-RP-Ink

Configuration

The following YAML configuration was used to produce this model:

models:
- model: EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2
- model: Sao10K/72B-Qwen2.5-Kunou-v1
- model: anthracite-org/magnum-v4-72b
- model: allura-org/Qwen2.5-72b-RP-Ink
merge_method: model_stock
base_model: migtissera/Tess-v2.5.2-Qwen2-72B
parameters:
  filter_wise: false
dytpe: float16
name: DatToad/Chuluun-Qwen2.5-72B-v0.08

Thank Yous!

Credit as always to the people who make the finetunes that go into this - they do the hard work, I just throw them in the blender! Most of them have Ko-fis, training isn't cheap and their time is valuable too. Special thanks to these contributors:

  • Everyone in Allura-Org and friends in Discord, for the EVA and Ink models, as well as all their support and mentoring that gave me the knowledge that makes merges like this possible.

  • Testers Geeechan and CURSE for invaluable feedback especially on optimal settings

  • Quant support from scene legends Bartowski and MikeRoz

  • All of you that have encouraged me and sent thanks and appreciation for this work. It wouldn't mean very much to me to keep this to myself.

Downloads last month
1
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MikeRoz/DatToad_Chuluun-Qwen2.5-72B-v0.08-4.25bpw-h6-exl2