Chuluun-Qwen2.5-72B-v0.08
This is a merge of pre-trained language models created using mergekit.
I re-ran the original Chuluun formula including the newly released Ink from Allura-Org. I've found the addition gives the model a lot more variability, likely because of aggressive de-slop applied to its dataset. Sometimes this means a word choice will be strange and you'll want to manually edit when needed, but it means you'll see less ministrations sparkling with mischief.
Because of this the best way to approach the model is to run multiple regens and choose the one you like, edit mercilessly, and continue. Like the original Chuluun this variant is very steerable for complex storywriting and RP. It's probably also a little spicier than v0.01 with both Magnum and whatever the heck Fizz threw into the data for Ink.
I've also been hearing praise for a level of character intelligence not seen in other models, including Largestral finetunes and merges. I'm not about to say any model of mine is smarter because it was a dumb idea to use Tess as the base and it somehow worked.
Tips for effective use
As with all writing-focused models balancing intelligence with creativity is tricky. If this one seems like it understands some details but not others try v0.01, overall I think this model is more creative but a little less coherent. If v0.08 is a little too chaotic for your tastes consider running v0.01 at first and switch to this model mid-story if it gets stale.
All the models within the merge use ChatML format so you'll want to use it too. Use Konnect's Qwenception prompt or whatever you prefer, it seems to do fine with any decent sysprompt. Lower temps are suggested than with v0.01 because of Ink's de-slop dataset, testers reported anywhere between 1 and 1.2 as a baseline but plan to adjust to taste. Consider dynatemp for this model as well. If dialogue gets repetitive at all that's usually a sign you need more temp.
Chuluun is built off of uncensored models and testing has confirmed it will go to dark places. Your use of this model assumes you to understand this and that you're responsible for what you do with it.
Merge Details
Merge Method
This model was merged using the Model Stock merge method using migtissera/Tess-v2.5.2-Qwen2-72B as a base.
Models Merged
The following models were included in the merge:
- EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2
- Sao10K/72B-Qwen2.5-Kunou-v1
- anthracite-org/magnum-v4-72b
- allura-org/Qwen2.5-72b-RP-Ink
Configuration
The following YAML configuration was used to produce this model:
models:
- model: EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2
- model: Sao10K/72B-Qwen2.5-Kunou-v1
- model: anthracite-org/magnum-v4-72b
- model: allura-org/Qwen2.5-72b-RP-Ink
merge_method: model_stock
base_model: migtissera/Tess-v2.5.2-Qwen2-72B
parameters:
filter_wise: false
dytpe: float16
name: DatToad/Chuluun-Qwen2.5-72B-v0.08
Thank Yous!
Credit as always to the people who make the finetunes that go into this - they do the hard work, I just throw them in the blender! Most of them have Ko-fis, training isn't cheap and their time is valuable too. Special thanks to these contributors:
Everyone in Allura-Org and friends in Discord, for the EVA and Ink models, as well as all their support and mentoring that gave me the knowledge that makes merges like this possible.
Testers Geechan and CURSE for invaluable feedback especially on optimal settings
Quant support from scene legends Bartowski and MikeRoz
All of you that have encouraged me and sent thanks and appreciation for this work. It wouldn't mean very much to me to keep this to myself.
- Downloads last month
- 14