--- tags: - roleplay - creative-writing - merge - mergekit base_model: - Delta-Vector/Hamanasu-Magnum-QwQ-32B - Sao10K/32B-Qwen2.5-Kunou-v1 pipeline_tag: text-generation library_name: transformers --- ``` __~a~_ ~~; ~_ _ ~ ~_ _ '_\;__._._._._._._] ~_._._._._._.__;/_` '(/'/'/'/'|'|'|'| ( )|'|'|'|'\'\'\'\)' (/ / / /, | | | |(/ \) | | | ,\ \ \ \) (/ / / / / | | | ~(/ \) ~ | | \ \ \ \ \) (/ / / / / ~ ~ ~ (/ \) ~ ~ \ \ \ \ \) (/ / / / ~ / (||)| ~ \ \ \ \) ~ / / ~ M /||\M ~ \ \ ~ ~ ~ /||\ ~ ~ //||\\ //||\\ //||\\ '/||\' "Archaeopteryx" ```

THESE ARE EXL2 QUANTS, LOOK IN THE REVISIONS FOR THE QUANTS, MAIN BRANCH CONTAINS MEASUREMENT.

A series of Merges made for Roleplaying & Creative Writing, This model uses 32B-Qwen2.5-Kunou-v1 and Hamanasu-Magnum-QwQ-32B and Slerp to merge the 2 models.

ChatML formatting

"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""
        

MergeKit Configuration

models:
  - model: Sao10K/32B-Qwen2.5-Kunou-v1
  - model: Delta-Vector/Hamanasu-Magnum-QwQ-32B
merge_method: slerp
base_model: Delta-Vector/Hamanasu-Magnum-QwQ-32B
parameters:
  t:
   - value: 0.2
dtype: bfloat16
tokenizer_source: base
        

Quants:

Credits

Thank you to: Kubernetes-bad, LucyKnada, Intervitens, Samantha Twinkman, Tav, Trappu & The rest of Anthracite