Nohobby's picture
Update README.md
47e1ce7 verified
metadata
base_model:
  - nbeerbower/Llama-3.1-Nemotron-lorablated-70B
  - sophosympatheia/New-Dawn-Llama-3.1-70B-v1.1
  - ArliAI/Llama-3.3-70B-ArliAI-RPMax-v1.4
  - SicariusSicariiStuff/Negative_LLAMA_70B
  - Sao10K/70B-L3.3-Cirrus-x1
  - Sao10K/L3.3-70B-Euryale-v2.3
  - Sao10K/L3.1-70B-Hanami-x1
  - TheDrummer/Anubis-70B-v1
  - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
  - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.0
  - Undi95/Sushi-v1.4
  - pankajmathur/orca_mini_v9_3_70B
  - unsloth/Llama-3.3-70B-Instruct
  - Blackroot/Mirai-3.0-70B
  - Black-Ink-Guild/Pernicious_Prophecy_70B
  - flammenai/Llama3.1-Flammades-70B
  - huihui-ai/DeepSeek-R1-Distill-Llama-70B-abliterated
  - Doctor-Shotgun/L3.3-70B-Magnum-v4-SE
  - Nexesenex/Llama_3.x_70b_Smarteaz_V1
  - crestf411/L3.1-nemotron-sunfall-v0.7.0
  - Fizzarolli/L3.1-70b-glitz-v0.2
library_name: transformers
tags:
  - mergekit
  - merge

Prikol

I don't even know anymore

Меня нужно изолировать от общества

Overview

After banging my head against the wall some more - I actually managed to merge DeepSeek distill into my mess! Along with even more models (my hand just slipped, I swear)

The prose is better than in v0.5, but has a different feel to it, so I guess it's more of a step to the side than forward (hence the title EXTRA instead of 0.6).

The context recall may have improved, or I'm just gaslighting myself to think so.

And of course, since it now has DeepSeek in it - <think> tags!

They kinda work out of the box if you add <think> to the 'Start Reply With' field in ST - that way the model will write a really short character thought in it. However, if we want some OOC reasoning, things get trickier.

My initial thought was that this model could be instructed to use <think> either only for {{char}}'s inner monologue or for detached analysis, but actually it would end up writing character thoughts most of the time anyway, and the times when it did reason stuff it threw the narrative out of the window by making it too formal and even adding some notes at the end.

And so the solution was to add a prefill after the <think> tag. There's a lot of room for improvement, but for now, I think this boats the float or whatever:

<think> [Okay, let me think through what's happening from my "morally ambiguous narrator" perspective before I continue this fictional roleplaying session]
*So

If you add the line break after the tag, the output becomes too formal, and if you remove the asterisk, it becomes too censored. Yeah...

Settings:

Prompt format: Llama3

Samplers: 1.2 Temp, 0.025 minP, 0.25 smoothing factor, 2.0 smoothing curve

Quants

Imatrix | Static

Merge Details

The things that I have done to bring about this abomination in our world are truly atrocious - as if v0.5 wasn't bad enough. Merging shouldn't be done the way I did it, really. Maybe one day I will bother to put out a branching diagram of this thing, since just listing the merge steps one by one is confusing.