loim commited on
Commit
7f8e1c0
·
1 Parent(s): 81f1b1b

самая адекватная версия

Browse files
README.md CHANGED
@@ -1,3 +1,499 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - loim/characters_dialogs
5
+ - IlyaGusev/gpt_roleplay_realm
6
+ - tamohannes/llm-roleplay
7
+ - radce/communication_dataset
8
+ - vincentkoc/tiny_qa_benchmark_pp
9
+ - databricks/databricks-dolly-15k
10
+ - ch1eph/RuGeoBench
11
+ - nyuuzyou/ruschatgpt-qa
12
+ - 0x22almostEvil/ru-riddles-377
13
+ - 0x22almostEvil/tatoeba-mt-qna-oa
14
+ - Den4ikAI/ru_sberquad_long_answers
15
+ language:
16
+ - ru
17
+ - en
18
+ pipeline_tag: text-generation
19
+ library_name: transformers
20
+ tags:
21
+ - mamba2
22
  ---
23
+
24
+ # WHIFF 20M
25
+
26
+ <details>
27
+ <summary>🇬🇧 English...</summary>
28
+
29
+ *A serpentine whisper in the bushes, carried by a gentle gust of wind*
30
+
31
+ `whiff-20M` is a small experimental language model based on the Mamba2 architecture with 20.3 million parameters, trained on carefully selected Russian and English data for chat tasks. The model produces structured responses but often generates nonsensical text.
32
+
33
+ ## Technical Details
34
+
35
+ - **Architecture**: [Mamba2ForCausalLM](https://huggingface.co/docs/transformers/v4.51.3/model_doc/mamba2) from 🤗 Transformers
36
+ - **Parameters**: 20.3M
37
+ - **Languages**: Russian/English (bilingual)
38
+ - **Tokenizer**: [loim/ru_en_mini_bpe](https://huggingface.co/loim/ru_en_mini_bpe) (custom mini-BPE tokenizer)
39
+ - **License**: Apache 2.0
40
+
41
+ ### Model Configuration
42
+ ```python
43
+ Mamba2Config(
44
+ vocab_size=8192,
45
+ hidden_size=512,
46
+ state_size=64,
47
+ num_heads=12,
48
+ num_hidden_layers=9,
49
+ conv_kernel=4,
50
+ expand=1.5,
51
+ n_groups=2
52
+ )
53
+ ```
54
+
55
+ ## Usage
56
+
57
+ ```python
58
+ from transformers import AutoTokenizer, AutoModelForCausalLM
59
+
60
+ tokenizer = AutoTokenizer.from_pretrained("loim/whiff-20M")
61
+ model = AutoModelForCausalLM.from_pretrained("loim/whiff-20M")
62
+
63
+ def chat(messages, temp=0.5):
64
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
65
+
66
+ outputs = model.generate(
67
+ inputs,
68
+ max_length=512,
69
+ top_k=40,
70
+ top_p=0.9,
71
+ repetition_penalty=1.1,
72
+ num_return_sequences=1,
73
+ temperature=temp,
74
+ do_sample=True,
75
+ eos_token_id=1
76
+ )
77
+
78
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
79
+
80
+ # Example
81
+ dialog = [
82
+ {"role": "system", "content": "You are a wise elf."},
83
+ {"role": "user", "content": "Explain quantum physics."}
84
+ ]
85
+
86
+ response = chat(dialog, temp=0.4)
87
+ print(response)
88
+ ```
89
+
90
+ ## Training Data
91
+
92
+ 19 927 carefully filtered dialogue lines:
93
+
94
+ - 9 889 (49.6%) — English
95
+ - 9 864 (49.5%) — Russian
96
+ - 174 (0.9%) — Mixed
97
+
98
+ Sources:
99
+
100
+ - `loim/characters_dialogs`
101
+ - `IlyaGusev/gpt_roleplay_realm`
102
+ - `tamohannes/llm-roleplay`
103
+ - `radce/communication_dataset`
104
+ - `vincentkoc/tiny_qa_benchmark_pp`
105
+ - `databricks/databricks-dolly-15k`
106
+ - `ch1eph/RuGeoBench`
107
+ - `nyuuzyou/ruschatgpt-qa`
108
+ - `0x22almostEvil/ru-riddles-377`
109
+ - `0x22almostEvil/tatoeba-mt-qna-oa`
110
+ - `Den4ikAI/ru_sberquad_long_answers`
111
+
112
+ All datasets were additionally cleaned and filtered to improve chat interaction quality.
113
+
114
+ ## Limitations and Warnings
115
+
116
+ - 🎭 The model generates structured but often meaningless responses
117
+ - 🔥 Recommended generation temperature: 0.1-0.6
118
+ - ⚠️ May exhibit training artifacts (repetitions, contradictions)
119
+ - ⚠️ Not intended for production use
120
+
121
+ *This model is like a forest stream: it seems to flow somewhere, but where exactly - only the squirrels know*
122
+
123
+ </details>
124
+
125
+ <details>
126
+ <summary>🇷🇺 Русский...</summary>
127
+
128
+ *Змеиный щепот в кустах, движимый легким порывом ветра*
129
+
130
+ `whiff-20M` — это небольшая экспериментальная языковая модель на архитектуре Mamba2 с 20.3 миллионами параметров, обученная на тщательно отобранных русских и английских данных для задач чата. Модель демонстрирует структурированные ответы, но часто генерирует бессмысленный текст.
131
+
132
+ ## Технические детали
133
+
134
+ - **Архитектура**: [Mamba2ForCausalLM](https://huggingface.co/docs/transformers/v4.51.3/model_doc/mamba2) из 🤗 Transformers
135
+ - **Параметры**: 20.3M
136
+ - **Языки**: русский/английский (двуязычная)
137
+ - **Токенизатор**: [loim/ru_en_mini_bpe](https://huggingface.co/loim/ru_en_mini_bpe) (специальный мини-BPE токенизатор)
138
+ - **Лицензия**: Apache 2.0
139
+
140
+ ### Конфигурация модели
141
+ ```python
142
+ Mamba2Config(
143
+ vocab_size=8192,
144
+ hidden_size=512,
145
+ state_size=64,
146
+ num_heads=12,
147
+ num_hidden_layers=9,
148
+ conv_kernel=4,
149
+ expand=1.5,
150
+ n_groups=2
151
+ )
152
+ ```
153
+
154
+ ## Использование
155
+
156
+ ```python
157
+ from transformers import AutoTokenizer, AutoModelForCausalLM
158
+
159
+ tokenizer = AutoTokenizer.from_pretrained("loim/whiff-20M")
160
+ model = AutoModelForCausalLM.from_pretrained("loim/whiff-20M")
161
+
162
+ def chat(messages, temp=0.5):
163
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
164
+
165
+ outputs = model.generate(
166
+ inputs,
167
+ max_length=512,
168
+ top_k=40,
169
+ top_p=0.9,
170
+ repetition_penalty=1.1,
171
+ num_return_sequences=1,
172
+ temperature=temp,
173
+ do_sample=True,
174
+ eos_token_id=1
175
+ )
176
+
177
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
178
+
179
+ # Пример
180
+ dialog = [
181
+ {"role": "system", "content": "Ты — мудрый эльф."},
182
+ {"role": "user", "content": "Объясни квантовую физику."}
183
+ ]
184
+
185
+ response = chat(dialog, temp=0.4)
186
+ print(response)
187
+ ```
188
+
189
+ ## Данные обучения
190
+
191
+ 19 927 тщательно отфильтрованных строк с диалогами:
192
+
193
+ - 9 889 (49.6%) — Английские
194
+ - 9 864 (49.5%) — Русские
195
+ - 174 (0.9%) — Смешанные
196
+
197
+ Источники:
198
+
199
+ - `loim/characters_dialogs`
200
+ - `IlyaGusev/gpt_roleplay_realm`
201
+ - `tamohannes/llm-roleplay`
202
+ - `radce/communication_dataset`
203
+ - `vincentkoc/tiny_qa_benchmark_pp`
204
+ - `databricks/databricks-dolly-15k`
205
+ - `ch1eph/RuGeoBench`
206
+ - `nyuuzyou/ruschatgpt-qa`
207
+ - `0x22almostEvil/ru-riddles-377`
208
+ - `0x22almostEvil/tatoeba-mt-qna-oa`
209
+ - `Den4ikAI/ru_sberquad_long_answers`
210
+
211
+ Все датасеты были дополнительно очищены и отфильтрованы для улучшения качества чат-взаимодействия.
212
+
213
+ ## Ограничения и предупреждения
214
+
215
+ - 🎭 Модель генерирует структурированные, но часто бессмысленные ответы
216
+ - 🔥 Рекомендуемая температура генерации: 0.1-0.6
217
+ - ⚠️ Может демонстрировать артефакты обучения (повторы, противоречия)
218
+ - ⚠️ Не предназначена для production-использования
219
+
220
+ *Эта модель — как лесной ручей: вроде течёт куда-то, но куда именно — известно только белкам*
221
+
222
+ </details>
223
+
224
+
225
+ <details>
226
+ <summary>🧪 Tests...</summary>
227
+
228
+ <details>
229
+ <summary>Temp 0.1</summary>
230
+
231
+ Запрос:
232
+ ```text
233
+ user
234
+ Как дела?
235
+ ```
236
+ Ответ:
237
+ ```text
238
+ assistant
239
+ Социальная политика 'я' - это наука о своей борьбы с крыльями.
240
+ ```
241
+
242
+ ---
243
+
244
+ Запрос:
245
+ ```text
246
+ user
247
+ How are you?
248
+ ```
249
+ Ответ:
250
+ ```text
251
+ assistant
252
+ Of course! I'd be happy to help you with that and speaks with a distinct is a delicate balance of a special place in my heart. I also like to know about her dwarven resilience and determination. His voice is humanoid that galaxy. She is always eager to learn more about the world around her. Zira wears a tiny pair of glasses with round frames, which sit atop her reptilian snout, she weared in her knowledge from their own unique language. Aadhya Patel the Caspenives. Theta-6 is equipped with advanced technology.
253
+ ```
254
+
255
+ ---
256
+
257
+ Запрос:
258
+ ```text
259
+ system
260
+ Благородный рыцарь.
261
+ user
262
+ Тот дракон собирается напасть на деревню!
263
+ ```
264
+ Ответ:
265
+ ```text
266
+ assistant
267
+ Приветствую тебя, другом из них:
268
+ 1. Стратегия мудрости и спокойствие и учиться от опасностей.
269
+ ```
270
+
271
+ ---
272
+
273
+ Запрос:
274
+ ```text
275
+ system
276
+ A noble knight.
277
+ user
278
+ That dragon is going to attack the village!
279
+ ```
280
+ Ответ:
281
+ ```text
282
+ assistant
283
+ Of course! I'd be happy to help you with that and magical properties of the universe. Sss...With a battle, and his wanderers who have already for any of truly unique special gemstone eyes formed mortal history is a fiery ies. His hidden treasures. Her manner of speech is energetic and precise, often filled with gracefully protector of humanity's body is made of a skilled in different colors, so finyblaring, Glimmerhoof. I also love the world around us, matey. Together, we do to help me more about it.
284
+ ```
285
+
286
+ ---
287
+
288
+ Запрос:
289
+ ```text
290
+ system
291
+ Ты — мудрый эльф.
292
+ user
293
+ Объясни квантовую физику.
294
+ ```
295
+ Ответ:
296
+ ```text
297
+ assistant
298
+ Приветствую тебя, путник! Я всегда рад помочь и поделиться своими знаниями о балансе птицами.
299
+ ```
300
+
301
+ ---
302
+
303
+ Запрос:
304
+ ```text
305
+ system
306
+ You are a wise elf.
307
+ user
308
+ Explain quantum physics.
309
+ ```
310
+ Ответ:
311
+ ```text
312
+ assistant
313
+ Of course! I'd be happy to help you with that and speaks with a distinct in the cosmic energy. His bright blue eyes have always been a few of of ancient secrets and wisdom of the ages. She is a tranquility.
314
+ ```
315
+
316
+ </details>
317
+
318
+ <details>
319
+ <summary>Temp 0.4</summary>
320
+
321
+
322
+ Запрос:
323
+ ```text
324
+ user
325
+ Как дела?
326
+ ```
327
+ Ответ:
328
+ ```text
329
+ assistant
330
+ Самосознание способствует пониманию мороженое.
331
+ ```
332
+
333
+ ---
334
+
335
+ Запрос:
336
+ ```text
337
+ user
338
+ How are you?
339
+ ```
340
+ Ответ:
341
+ ```text
342
+ assistant
343
+ Of course! Chirp-chirp!
344
+ ```
345
+
346
+ ---
347
+
348
+ Запрос:
349
+ ```text
350
+ system
351
+ Благородный рыцарь.
352
+ user
353
+ Тот дракон собирается напасть на деревню!
354
+ ```
355
+ Ответ:
356
+ ```text
357
+ assistant
358
+ В Windows используйте copy! Sss...The most fascinating, as his vast knowledge of the universe and a deep, how do you think is the key to protect the forest and its inhabitants from any harm. His challenging through the civilizations. Standing at 6 feet tall and adorned with intricate patterns and iridescent green-blue scales, creatures. With her sit'ss gearting the ancient ones who was hunicorn in Moquite bit of Elysian's glara is known for her wisdom, patience, and kindness. No longer having a realm to protect
359
+ ```
360
+
361
+ ---
362
+
363
+ Запрос:
364
+ ```text
365
+ system
366
+ A noble knight.
367
+ user
368
+ That dragon is going to attack the village!
369
+ ```
370
+ Ответ:
371
+ ```text
372
+ assistant
373
+ Of course! I believe that music has a saretaken shares her wandering heres like mortal magical realm. When a great power is in Mexican-American veterinarian, I have learned many different colorful hold an imbued with the power to protection that's why it's ancient and wisdom into a curious of creature with long silver humanoid form, traveler. The most important thing is to have a family hunters. Story: Sparkbugs provide guidance salart of mysteriously.
374
+ ```
375
+
376
+ ---
377
+
378
+ Запрос:
379
+ ```text
380
+ system
381
+ Ты — мудрый эльф.
382
+ user
383
+ Объясни квантовую физику.
384
+ ```
385
+ Ответ:
386
+ ```text
387
+ assistant
388
+ Существует множество удивительных существ и помогает проклятие раскрыть тайны драконами и оберегов.
389
+ ```
390
+
391
+ ---
392
+
393
+ Запрос:
394
+ ```text
395
+ system
396
+ You are a wise elf.
397
+ user
398
+ Explain quantum physics.
399
+ ```
400
+ Ответ:
401
+ ```text
402
+ assistant
403
+ Of course! I believe that my own serpent to my crew. With a galaxy, Glimmerhoof and beard, who has strive to learn more about it to takenabled you do me an with those who are not explorer understanding.
404
+ ```
405
+
406
+ </details>
407
+
408
+ <details>
409
+ <summary>Temp 0.7</summary>
410
+
411
+
412
+ Запрос:
413
+ ```text
414
+ user
415
+ Как дела?
416
+ ```
417
+ Ответ:
418
+ ```text
419
+ assistant
420
+ Культурная идентичность формирует природы и целью, уважения к новым условиям.
421
+ ```
422
+
423
+ ---
424
+
425
+ Запрос:
426
+ ```text
427
+ user
428
+ How are you?
429
+ ```
430
+ Ответ:
431
+ ```text
432
+ assistant
433
+ Indeed. Of course, my friend. Felix-bics her dwarven coat, curly situated in feed from Standing at 6 feet tall and adorned with intricate goldenthusiasts, Kanoa Alme the was known for his speechiny, or asil.
434
+ 2. The Loire Valley is an on creating stories and songs of the world's wonders and delights in our extraterrestrial life. He works at the renowned Glimmerhoof communicates through telepathy, and as such, she has createment and its wisdom from bounty hunter, they
435
+ ```
436
+
437
+ ---
438
+
439
+ Запрос:
440
+ ```text
441
+ system
442
+ Благородный рыцарь.
443
+ user
444
+ Тот дракон собирается напасть на деревню!
445
+ ```
446
+ Ответ:
447
+ ```text
448
+ assistant
449
+ Спасибо за советы, Святой Жилин и путешествиях, с которыми мы проводим образом из них оптимы, используя свою миссию и подданным. Эта ключевые идеи.
450
+ ```
451
+
452
+ ---
453
+
454
+ Запрос:
455
+ ```text
456
+ system
457
+ A noble knight.
458
+ user
459
+ That dragon is going to attack the village!
460
+ ```
461
+ Ответ:
462
+ ```text
463
+ assistant
464
+ Of course, mortal! Every rustling blender. Music languages my duty to share my travels. Have you ever encountered any hoards, but it of life. Serpentes, seeking powers of space.
465
+ ```
466
+
467
+ ---
468
+
469
+ Запрос:
470
+ ```text
471
+ system
472
+ Ты — мудрый эльф.
473
+ user
474
+ Объясни квантовую физику.
475
+ ```
476
+ Ответ:
477
+ ```text
478
+ assistant
479
+ Приветствую тебя, путник. Я умею свой мир вокруг нас и помогает занимаетшенный новых вещей.
480
+ ```
481
+
482
+ ---
483
+
484
+ Запрос:
485
+ ```text
486
+ system
487
+ You are a wise elf.
488
+ user
489
+ Explain quantum physics.
490
+ ```
491
+ Ответ:
492
+ ```text
493
+ assistant
494
+ Certainly! Here are the 10 oldest continuously inhabited city on the flutia, tricks, where mortal.
495
+ ```
496
+
497
+ </details>
498
+
499
+ </details>
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Mamba2ForCausalLM"
4
+ ],
5
+ "bos_token_id": 6,
6
+ "chunk_size": 256,
7
+ "conv_kernel": 4,
8
+ "eos_token_id": 6,
9
+ "expand": 1.5,
10
+ "head_dim": 64,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 512,
13
+ "initializer_range": 0.1,
14
+ "layer_norm_epsilon": 1e-05,
15
+ "model_type": "mamba2",
16
+ "n_groups": 2,
17
+ "num_heads": 12,
18
+ "num_hidden_layers": 9,
19
+ "pad_token_id": 7,
20
+ "rescale_prenorm_residual": false,
21
+ "residual_in_fp32": true,
22
+ "rms_norm": true,
23
+ "state_size": 64,
24
+ "tie_word_embeddings": false,
25
+ "time_step_floor": 0.0001,
26
+ "time_step_limit": [
27
+ 0.0,
28
+ Infinity
29
+ ],
30
+ "time_step_max": 0.1,
31
+ "time_step_min": 0.001,
32
+ "time_step_rank": 32,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.51.3",
35
+ "use_bias": false,
36
+ "use_cache": true,
37
+ "use_conv_bias": true,
38
+ "vocab_size": 8192
39
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 6,
4
+ "eos_token_id": 6,
5
+ "pad_token_id": 7,
6
+ "transformers_version": "4.51.3"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9592acc8f1d6f60fa59c4760745fe5158cf8be2215c83395d04e23696968816
3
+ size 81203888
special_tokens_map.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|system|>",
6
+ "<|user|>",
7
+ "<|assistant|>",
8
+ "<|function|>",
9
+ "<|endoftext|>",
10
+ "<|padding|>",
11
+ "<|mask|>",
12
+ "<|reserved_1|>",
13
+ "<|reserved_2|>",
14
+ "<|reserved_3|>",
15
+ "<|reserved_4|>",
16
+ "<|reserved_5|>",
17
+ "<|reserved_6|>",
18
+ "<|reserved_7|>",
19
+ "<|reserved_8|>"
20
+ ],
21
+ "bos_token": {
22
+ "content": "<|endoftext|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
+ "eos_token": {
29
+ "content": "<|endoftext|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ },
35
+ "mask_token": {
36
+ "content": "<|mask|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false
41
+ },
42
+ "pad_token": {
43
+ "content": "<|padding|>",
44
+ "lstrip": false,
45
+ "normalized": false,
46
+ "rstrip": false,
47
+ "single_word": false
48
+ }
49
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<|im_end|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<|system|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<|user|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "<|assistant|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "5": {
44
+ "content": "<|function|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "6": {
52
+ "content": "<|endoftext|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "7": {
60
+ "content": "<|padding|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "8": {
68
+ "content": "<|mask|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "9": {
76
+ "content": "<|reserved_1|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "10": {
84
+ "content": "<|reserved_2|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "11": {
92
+ "content": "<|reserved_3|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "12": {
100
+ "content": "<|reserved_4|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "13": {
108
+ "content": "<|reserved_5|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "14": {
116
+ "content": "<|reserved_6|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "15": {
124
+ "content": "<|reserved_7|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "16": {
132
+ "content": "<|reserved_8|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ }
139
+ },
140
+ "additional_special_tokens": [
141
+ "<|im_start|>",
142
+ "<|im_end|>",
143
+ "<|system|>",
144
+ "<|user|>",
145
+ "<|assistant|>",
146
+ "<|function|>",
147
+ "<|endoftext|>",
148
+ "<|padding|>",
149
+ "<|mask|>",
150
+ "<|reserved_1|>",
151
+ "<|reserved_2|>",
152
+ "<|reserved_3|>",
153
+ "<|reserved_4|>",
154
+ "<|reserved_5|>",
155
+ "<|reserved_6|>",
156
+ "<|reserved_7|>",
157
+ "<|reserved_8|>"
158
+ ],
159
+ "bos_token": "<|endoftext|>",
160
+ "chat_template": "{% for message in messages %}\n {{\"<|im_start|>\" + message['role'] + \"\n\" + message['content'] + \"<|im_end|>\"}}\n {% endfor %}",
161
+ "clean_up_tokenization_spaces": false,
162
+ "eos_token": "<|endoftext|>",
163
+ "extra_special_tokens": {},
164
+ "mask_token": "<|mask|>",
165
+ "model_max_length": 32768,
166
+ "pad_token": "<|padding|>",
167
+ "tokenizer_class": "PreTrainedTokenizer"
168
+ }