Text2Text Generation
Safetensors
English
gemma3
Not-For-All-Audiences
SicariusSicariiStuff commited on
Commit
74a8dd7
·
verified ·
1 Parent(s): 059ce42

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +276 -0
README.md ADDED
@@ -0,0 +1,276 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: gemma
3
+ language:
4
+ - en
5
+ base_model:
6
+ - google/gemma-3-12b-it
7
+ tags:
8
+ - not-for-all-audiences
9
+ pipeline_tag: image-text-to-text
10
+ ---
11
+
12
+ <div align="center">
13
+ <b style="font-size: 40px;">Oni_Mitsubishi_12B</b>
14
+
15
+
16
+ </div>
17
+
18
+
19
+ <img src="https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B/resolve/main/Images/Oni_Mitsubishi_12B.png" alt="Oni_Mitsubishi_12B" style="width: 70%; min-width: 500px; display: block; margin: auto;">
20
+
21
+
22
+ ---
23
+
24
+ <a href="https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B#tldr" style="color: purple; font-weight: bold; font-size: 48px; text-decoration: none; display: block; text-align: center;">Click here for TL;DR</a>
25
+
26
+ ---
27
+ **It happened**. The long-awaited **Gemma-3** is here, and not only are the model sizes really good (**1, 4, 12, 27**), but the **128k** context (except for the 1B 32k) was exactly what the Open-Source community wanted and asked for. My only issue with Gemma models in general, is the VRAM requirement for **tuning them**, but that's a "me problem." End users will probably be very happy with Gemma-3 in terms of the VRAM requirement for **running it**.
28
+
29
+ On the **12th** of March, the Gemma-3 family of models was released. So I decided to go **full superstitious**, and took this omen as a divine calling to finetune the **12B** model first. This is how **Oni_Mitsubishi_12B** was born.
30
+
31
+ Before starting the actual training run, I used the following command, which I believe has helped the model to converge "better":
32
+ ```
33
+ for i in {1..666}; do nvidia-smi; done
34
+ ```
35
+
36
+
37
+ Gemma is known for its "**Gemma knowledge**": fandom and \ or other obscure knowledge that sometimes even larger LLMs often do not possess. It gets even better, as this time we also got a **vision model** embedded into all the Gemma-3 models, except for the 1B. I wonder what are the possibilities for the vision part if the text layers are uncensored?
38
+
39
+ I have used brand new **long context markdown data**, some **deslopped** instruct data (very lightly deslopped, it's very time-consuming to get right), **and more than 50%** of highly curated and filtered organic human data, meticulously cleaned, and parsed into obedience. A new stack of organic and data-engineered text was used **for the first time** for **Oni_Mitsubishi_12B**. I truly hope creating it was worth the effort.
40
+
41
+ At **NO POINT** ChatGPT was used for data generation, however, the new **Claude 3.7** sonnet was used **VERY** sparingly for the **specific task** of creating a small number of humorous datasets (very human-like, was done with a decent amount of prompt engineering), I've meticulously checked them for slop, and it is **minimal**. This goal of said data was to imitate human text, using the **4chan vernacular**.
42
+
43
+ Speaking of which, I've published a highly curated, SFT-ready 4chan dataset here: [UBW_Tapestries](https://huggingface.co/datasets/SicariusSicariiStuff/UBW_Tapestries), naturally I have included it in the dataset used for this model as well.
44
+
45
+ ---
46
+
47
+ # Technical details
48
+
49
+ I've used the "ancient" **Alpaca chat template** because the **Gemma-3 chat template** was behaving funkily, and I didn't want to waste precious time, and instead give the community a more uncensored finetune to play with, as fast as possible (I saw this requested a lot on both Reddit and discord, understandable). In my opinion, it's silly to let perfect be an enemy of the good. Anyway, I had to use both bleeding edge **Transformers** and **Axolotl**, and modify stuff **that wasn't even supposed to work** (like the model's config.json).
50
+
51
+ Since it's a hybrid model, training its text-only part is a bit problematic, so I hacked a config.json that gaslights the model into thinking it's only a text model, and got some warnings like:
52
+
53
+ ```
54
+ 'vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.weight', 'vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias'}
55
+ - This IS expected if you are initializing Gemma3ForCausalLM from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
56
+ - This IS NOT expected if you are initializing Gemma3ForCausalLM from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
57
+ ```
58
+
59
+ Then I saw it trains.
60
+
61
+ >The absolute state when you can train a model before you can actually inference it.
62
+
63
+ ## Feedback, as always, is very much welcomed (even if it's negative).
64
+
65
+ ---
66
+
67
+ # Included Character cards in this repo:
68
+
69
+ - [Vesper](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B/resolve/main/Character_Cards/Takai_Puraisu.png) (Car dealership simulator)
70
+
71
+ ---
72
+
73
+ # Other character cards:
74
+ - [Vesper](https://huggingface.co/SicariusSicariiStuff/Phi-Line_14B/resolve/main/Character_Cards/Vesper.png) (Schizo **Space Adventure**)
75
+ - [Nina_Nakamura](https://huggingface.co/SicariusSicariiStuff/Phi-Line_14B/resolve/main/Character_Cards/Nina_Nakamura.png) (The **sweetest** dorky co-worker)
76
+ - [Employe#11](https://huggingface.co/SicariusSicariiStuff/Phi-Line_14B/resolve/main/Character_Cards/Employee%2311.png) (**Schizo workplace** with a **schizo worker**)
77
+
78
+ ---
79
+
80
+ ### TL;DR
81
+ - **Excellent Roleplay** abilities. Like Gemma-2, but better in every way.
82
+ - **Medium length** response (1-4 paragraphs, usually 2-3).
83
+ - **Excellent assistant** great with tables and markdown.
84
+ - Strong **Creative writing** abilities due to huge chunk of organic creative writing data. Will obey requests regarding formatting (markdown headlines for paragraphs, etc).
85
+ - **LOW refusals** - Total freedom in RP, can do things other RP models won't, and I'll leave it at that. Low refusals in assistant tasks as well.
86
+ - **VERY good** at following the **character card**. Based on the best RP datasets I have available.
87
+
88
+ ### Important: Make sure to use the correct settings!
89
+ [Assistant settings](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B#recommended-settings-for-assistant-mode)
90
+
91
+ [Roleplay settings](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B#recommended-settings-for-roleplay-mode)
92
+
93
+
94
+ ---
95
+
96
+ ## Oni_Mitsubishi_12B is available at the following quantizations:
97
+
98
+ - Original: [FP16](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B)
99
+ - GGUF & iMatrix: [GGUF](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B_GGUF) | [iMatrix](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B_iMatrix)
100
+ - Specialized: [FP8](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B_FP8)
101
+ - Mobile (ARM): [Q4_0](https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B_ARM)
102
+ ---
103
+
104
+ ## Model Details
105
+
106
+ - Intended use: **Role-Play**, **Creative Writing**, **General Tasks**.
107
+
108
+ - Censorship level: <b>Low - Medium</b>
109
+
110
+ - **X / 10** (10 completely uncensored)
111
+
112
+
113
+ ## UGI score:
114
+
115
+
116
+ Soon...
117
+
118
+
119
+
120
+ ---
121
+
122
+
123
+ ## Recommended settings for assistant mode
124
+ <details>
125
+ <summary>Full generation settings: <b>Debug Deterministic</b>.</summary>
126
+
127
+ <img src="https://huggingface.co/SicariusSicariiStuff/Dusk_Rainbow/resolve/main/Presets/Debug-deterministic.png" alt="Oni_Mitsubishi_12B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">
128
+
129
+ </details>
130
+
131
+ <details>
132
+ <summary>Full generation settings: <b>min_p</b>.</summary>
133
+
134
+ <img src="https://huggingface.co/SicariusSicariiStuff/Dusk_Rainbow/resolve/main/Presets/min_p.png" alt="Oni_Mitsubishi_12B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">
135
+
136
+ </details>
137
+
138
+ ---
139
+
140
+ <h2 style="color: green; font-weight: bold; font-size: 36px; text-align: center;">Settings for RP, click below to expand:</h2>
141
+
142
+ <details>
143
+ <summary><b>Roleplay settings:</b>.</summary>
144
+ A good repetition_penalty range is <b>between 1.12 - 1.15</b>, feel free to experiment.
145
+
146
+ With these settings, each output message should be neatly displayed in <b>1 - 5</b> paragraphs, <b>2 - 3</b> is the most common. A single paragraph will be output as a response to a simple message ("What was your name again?").
147
+
148
+ <b>min_P</b> for RP works too but is more likely to put everything under one large paragraph, instead of a neatly formatted short one. Feel free to switch in between.
149
+
150
+ <b>(Open the image in a new window to better see the full details)</b>
151
+ <img src="https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B/resolve/main/Presets/Oni_Mitsubishi_12B_RP.png" alt="Oni_Mitsubishi_12B_Settings" style="width: 100%; min-width: 600px; display: block; margin: auto;">
152
+
153
+ ```
154
+ temperature: 0.8
155
+ top_p: 0.95
156
+ top_k: 25
157
+ typical_p: 1
158
+ min_p: 0
159
+ repetition_penalty: 1.12
160
+ repetition_penalty_range: 1024
161
+ ```
162
+
163
+ </details>
164
+
165
+
166
+ <h2 style="color: darkorange; font-weight: bold; font-size: 65px; text-align: center;">Roleplay format: Classic Internet RP</h2>
167
+
168
+ ```
169
+ *action* speech *narration*
170
+ ```
171
+
172
+ - **min_p** will bias towards a **single big paragraph**.
173
+ - The recommended RP settings will bias towards **1-3 small paragraphs** (on some occasions 4-5)
174
+
175
+ ---
176
+
177
+
178
+
179
+ # Model instruction template: Alpaca
180
+
181
+ ```
182
+ ### Instruction:
183
+ {prompt}
184
+
185
+ ### Response:
186
+ ```
187
+
188
+ ---
189
+
190
+ **Other recommended generation Presets:**
191
+
192
+ <details>
193
+ <summary><b>Midnight Enigma</b></summary>
194
+
195
+ ```
196
+ max_new_tokens: 512
197
+ temperature: 0.98
198
+ top_p: 0.37
199
+ top_k: 100
200
+ typical_p: 1
201
+ min_p: 0
202
+ repetition_penalty: 1.18
203
+ do_sample: True
204
+ ```
205
+
206
+
207
+ </details>
208
+
209
+
210
+ <details>
211
+ <summary><b>Divine Intellect</b></summary>
212
+
213
+ ```
214
+ max_new_tokens: 512
215
+ temperature: 1.31
216
+ top_p: 0.14
217
+ top_k: 49
218
+ typical_p: 1
219
+ min_p: 0
220
+ repetition_penalty: 1.17
221
+ do_sample: True
222
+ ```
223
+
224
+
225
+ </details>
226
+
227
+ <details>
228
+ <summary><b>simple-1</b></summary>
229
+
230
+ ```
231
+ max_new_tokens: 512
232
+ temperature: 0.7
233
+ top_p: 0.9
234
+ top_k: 20
235
+ typical_p: 1
236
+ min_p: 0
237
+ repetition_penalty: 1.15
238
+ do_sample: True
239
+ ```
240
+
241
+
242
+ </details>
243
+
244
+ ---
245
+
246
+ <h2 style="color: green; font-weight: bold; font-size: 65px; text-align: center;">Your support = more models</h2>
247
+ <a href="https://ko-fi.com/sicarius" style="color: pink; font-weight: bold; font-size: 48px; text-decoration: none; display: block; text-align: center;">My Ko-fi page (Click here)</a>
248
+
249
+ ---
250
+
251
+
252
+ ## Citation Information
253
+
254
+ ```
255
+ @llm{Oni_Mitsubishi_12B,
256
+ author = {SicariusSicariiStuff},
257
+ title = {Oni_Mitsubishi_12B},
258
+ year = {2025},
259
+ publisher = {Hugging Face},
260
+ url = {https://huggingface.co/SicariusSicariiStuff/Oni_Mitsubishi_12B}
261
+ }
262
+ ```
263
+
264
+ ---
265
+
266
+ ## Benchmarks
267
+
268
+
269
+ Soon...
270
+
271
+ ---
272
+
273
+ ## Other stuff
274
+ - [SLOP_Detector](https://github.com/SicariusSicariiStuff/SLOP_Detector) Nuke GPTisms, with SLOP detector.
275
+ - [LLAMA-3_8B_Unaligned](https://huggingface.co/SicariusSicariiStuff/LLAMA-3_8B_Unaligned) The grand project that started it all.
276
+ - [Blog and updates (Archived)](https://huggingface.co/SicariusSicariiStuff/Blog_And_Updates) Some updates, some rambles, sort of a mix between a diary and a blog.