самая адекватная версия
Browse files- README.md +496 -0
- config.json +39 -0
- generation_config.json +7 -0
- model.safetensors +3 -0
- special_tokens_map.json +49 -0
- tokenizer.json +0 -0
- tokenizer_config.json +168 -0
README.md
CHANGED
@@ -1,3 +1,499 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- loim/characters_dialogs
|
5 |
+
- IlyaGusev/gpt_roleplay_realm
|
6 |
+
- tamohannes/llm-roleplay
|
7 |
+
- radce/communication_dataset
|
8 |
+
- vincentkoc/tiny_qa_benchmark_pp
|
9 |
+
- databricks/databricks-dolly-15k
|
10 |
+
- ch1eph/RuGeoBench
|
11 |
+
- nyuuzyou/ruschatgpt-qa
|
12 |
+
- 0x22almostEvil/ru-riddles-377
|
13 |
+
- 0x22almostEvil/tatoeba-mt-qna-oa
|
14 |
+
- Den4ikAI/ru_sberquad_long_answers
|
15 |
+
language:
|
16 |
+
- ru
|
17 |
+
- en
|
18 |
+
pipeline_tag: text-generation
|
19 |
+
library_name: transformers
|
20 |
+
tags:
|
21 |
+
- mamba2
|
22 |
---
|
23 |
+
|
24 |
+
# WHIFF 20M
|
25 |
+
|
26 |
+
<details>
|
27 |
+
<summary>🇬🇧 English...</summary>
|
28 |
+
|
29 |
+
*A serpentine whisper in the bushes, carried by a gentle gust of wind*
|
30 |
+
|
31 |
+
`whiff-20M` is a small experimental language model based on the Mamba2 architecture with 20.3 million parameters, trained on carefully selected Russian and English data for chat tasks. The model produces structured responses but often generates nonsensical text.
|
32 |
+
|
33 |
+
## Technical Details
|
34 |
+
|
35 |
+
- **Architecture**: [Mamba2ForCausalLM](https://huggingface.co/docs/transformers/v4.51.3/model_doc/mamba2) from 🤗 Transformers
|
36 |
+
- **Parameters**: 20.3M
|
37 |
+
- **Languages**: Russian/English (bilingual)
|
38 |
+
- **Tokenizer**: [loim/ru_en_mini_bpe](https://huggingface.co/loim/ru_en_mini_bpe) (custom mini-BPE tokenizer)
|
39 |
+
- **License**: Apache 2.0
|
40 |
+
|
41 |
+
### Model Configuration
|
42 |
+
```python
|
43 |
+
Mamba2Config(
|
44 |
+
vocab_size=8192,
|
45 |
+
hidden_size=512,
|
46 |
+
state_size=64,
|
47 |
+
num_heads=12,
|
48 |
+
num_hidden_layers=9,
|
49 |
+
conv_kernel=4,
|
50 |
+
expand=1.5,
|
51 |
+
n_groups=2
|
52 |
+
)
|
53 |
+
```
|
54 |
+
|
55 |
+
## Usage
|
56 |
+
|
57 |
+
```python
|
58 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
59 |
+
|
60 |
+
tokenizer = AutoTokenizer.from_pretrained("loim/whiff-20M")
|
61 |
+
model = AutoModelForCausalLM.from_pretrained("loim/whiff-20M")
|
62 |
+
|
63 |
+
def chat(messages, temp=0.5):
|
64 |
+
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
|
65 |
+
|
66 |
+
outputs = model.generate(
|
67 |
+
inputs,
|
68 |
+
max_length=512,
|
69 |
+
top_k=40,
|
70 |
+
top_p=0.9,
|
71 |
+
repetition_penalty=1.1,
|
72 |
+
num_return_sequences=1,
|
73 |
+
temperature=temp,
|
74 |
+
do_sample=True,
|
75 |
+
eos_token_id=1
|
76 |
+
)
|
77 |
+
|
78 |
+
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
79 |
+
|
80 |
+
# Example
|
81 |
+
dialog = [
|
82 |
+
{"role": "system", "content": "You are a wise elf."},
|
83 |
+
{"role": "user", "content": "Explain quantum physics."}
|
84 |
+
]
|
85 |
+
|
86 |
+
response = chat(dialog, temp=0.4)
|
87 |
+
print(response)
|
88 |
+
```
|
89 |
+
|
90 |
+
## Training Data
|
91 |
+
|
92 |
+
19 927 carefully filtered dialogue lines:
|
93 |
+
|
94 |
+
- 9 889 (49.6%) — English
|
95 |
+
- 9 864 (49.5%) — Russian
|
96 |
+
- 174 (0.9%) — Mixed
|
97 |
+
|
98 |
+
Sources:
|
99 |
+
|
100 |
+
- `loim/characters_dialogs`
|
101 |
+
- `IlyaGusev/gpt_roleplay_realm`
|
102 |
+
- `tamohannes/llm-roleplay`
|
103 |
+
- `radce/communication_dataset`
|
104 |
+
- `vincentkoc/tiny_qa_benchmark_pp`
|
105 |
+
- `databricks/databricks-dolly-15k`
|
106 |
+
- `ch1eph/RuGeoBench`
|
107 |
+
- `nyuuzyou/ruschatgpt-qa`
|
108 |
+
- `0x22almostEvil/ru-riddles-377`
|
109 |
+
- `0x22almostEvil/tatoeba-mt-qna-oa`
|
110 |
+
- `Den4ikAI/ru_sberquad_long_answers`
|
111 |
+
|
112 |
+
All datasets were additionally cleaned and filtered to improve chat interaction quality.
|
113 |
+
|
114 |
+
## Limitations and Warnings
|
115 |
+
|
116 |
+
- 🎭 The model generates structured but often meaningless responses
|
117 |
+
- 🔥 Recommended generation temperature: 0.1-0.6
|
118 |
+
- ⚠️ May exhibit training artifacts (repetitions, contradictions)
|
119 |
+
- ⚠️ Not intended for production use
|
120 |
+
|
121 |
+
*This model is like a forest stream: it seems to flow somewhere, but where exactly - only the squirrels know*
|
122 |
+
|
123 |
+
</details>
|
124 |
+
|
125 |
+
<details>
|
126 |
+
<summary>🇷🇺 Русский...</summary>
|
127 |
+
|
128 |
+
*Змеиный щепот в кустах, движимый легким порывом ветра*
|
129 |
+
|
130 |
+
`whiff-20M` — это небольшая экспериментальная языковая модель на архитектуре Mamba2 с 20.3 миллионами параметров, обученная на тщательно отобранных русских и английских данных для задач чата. Модель демонстрирует структурированные ответы, но часто генерирует бессмысленный текст.
|
131 |
+
|
132 |
+
## Технические детали
|
133 |
+
|
134 |
+
- **Архитектура**: [Mamba2ForCausalLM](https://huggingface.co/docs/transformers/v4.51.3/model_doc/mamba2) из 🤗 Transformers
|
135 |
+
- **Параметры**: 20.3M
|
136 |
+
- **Языки**: русский/английский (двуязычная)
|
137 |
+
- **Токенизатор**: [loim/ru_en_mini_bpe](https://huggingface.co/loim/ru_en_mini_bpe) (специальный мини-BPE токенизатор)
|
138 |
+
- **Лицензия**: Apache 2.0
|
139 |
+
|
140 |
+
### Конфигурация модели
|
141 |
+
```python
|
142 |
+
Mamba2Config(
|
143 |
+
vocab_size=8192,
|
144 |
+
hidden_size=512,
|
145 |
+
state_size=64,
|
146 |
+
num_heads=12,
|
147 |
+
num_hidden_layers=9,
|
148 |
+
conv_kernel=4,
|
149 |
+
expand=1.5,
|
150 |
+
n_groups=2
|
151 |
+
)
|
152 |
+
```
|
153 |
+
|
154 |
+
## Использование
|
155 |
+
|
156 |
+
```python
|
157 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
158 |
+
|
159 |
+
tokenizer = AutoTokenizer.from_pretrained("loim/whiff-20M")
|
160 |
+
model = AutoModelForCausalLM.from_pretrained("loim/whiff-20M")
|
161 |
+
|
162 |
+
def chat(messages, temp=0.5):
|
163 |
+
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
|
164 |
+
|
165 |
+
outputs = model.generate(
|
166 |
+
inputs,
|
167 |
+
max_length=512,
|
168 |
+
top_k=40,
|
169 |
+
top_p=0.9,
|
170 |
+
repetition_penalty=1.1,
|
171 |
+
num_return_sequences=1,
|
172 |
+
temperature=temp,
|
173 |
+
do_sample=True,
|
174 |
+
eos_token_id=1
|
175 |
+
)
|
176 |
+
|
177 |
+
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
178 |
+
|
179 |
+
# Пример
|
180 |
+
dialog = [
|
181 |
+
{"role": "system", "content": "Ты — мудрый эльф."},
|
182 |
+
{"role": "user", "content": "Объясни квантовую физику."}
|
183 |
+
]
|
184 |
+
|
185 |
+
response = chat(dialog, temp=0.4)
|
186 |
+
print(response)
|
187 |
+
```
|
188 |
+
|
189 |
+
## Данные обучения
|
190 |
+
|
191 |
+
19 927 тщательно отфильтрованных строк с диалогами:
|
192 |
+
|
193 |
+
- 9 889 (49.6%) — Английские
|
194 |
+
- 9 864 (49.5%) — Русские
|
195 |
+
- 174 (0.9%) — Смешанные
|
196 |
+
|
197 |
+
Источники:
|
198 |
+
|
199 |
+
- `loim/characters_dialogs`
|
200 |
+
- `IlyaGusev/gpt_roleplay_realm`
|
201 |
+
- `tamohannes/llm-roleplay`
|
202 |
+
- `radce/communication_dataset`
|
203 |
+
- `vincentkoc/tiny_qa_benchmark_pp`
|
204 |
+
- `databricks/databricks-dolly-15k`
|
205 |
+
- `ch1eph/RuGeoBench`
|
206 |
+
- `nyuuzyou/ruschatgpt-qa`
|
207 |
+
- `0x22almostEvil/ru-riddles-377`
|
208 |
+
- `0x22almostEvil/tatoeba-mt-qna-oa`
|
209 |
+
- `Den4ikAI/ru_sberquad_long_answers`
|
210 |
+
|
211 |
+
Все датасеты были дополнительно очищены и отфильтрованы для улучшения качества чат-взаимодействия.
|
212 |
+
|
213 |
+
## Ограничения и предупреждения
|
214 |
+
|
215 |
+
- 🎭 Модель генерирует структурированные, но часто бессмысленные ответы
|
216 |
+
- 🔥 Рекомендуемая температура генерации: 0.1-0.6
|
217 |
+
- ⚠️ Может демонстрировать артефакты обучения (повторы, противоречия)
|
218 |
+
- ⚠️ Не предназначена для production-использования
|
219 |
+
|
220 |
+
*Эта модель — как лесной ручей: вроде течёт куда-то, но куда именно — известно только белкам*
|
221 |
+
|
222 |
+
</details>
|
223 |
+
|
224 |
+
|
225 |
+
<details>
|
226 |
+
<summary>🧪 Tests...</summary>
|
227 |
+
|
228 |
+
<details>
|
229 |
+
<summary>Temp 0.1</summary>
|
230 |
+
|
231 |
+
Запрос:
|
232 |
+
```text
|
233 |
+
user
|
234 |
+
Как дела?
|
235 |
+
```
|
236 |
+
Ответ:
|
237 |
+
```text
|
238 |
+
assistant
|
239 |
+
Социальная политика 'я' - это наука о своей борьбы с крыльями.
|
240 |
+
```
|
241 |
+
|
242 |
+
---
|
243 |
+
|
244 |
+
Запрос:
|
245 |
+
```text
|
246 |
+
user
|
247 |
+
How are you?
|
248 |
+
```
|
249 |
+
Ответ:
|
250 |
+
```text
|
251 |
+
assistant
|
252 |
+
Of course! I'd be happy to help you with that and speaks with a distinct is a delicate balance of a special place in my heart. I also like to know about her dwarven resilience and determination. His voice is humanoid that galaxy. She is always eager to learn more about the world around her. Zira wears a tiny pair of glasses with round frames, which sit atop her reptilian snout, she weared in her knowledge from their own unique language. Aadhya Patel the Caspenives. Theta-6 is equipped with advanced technology.
|
253 |
+
```
|
254 |
+
|
255 |
+
---
|
256 |
+
|
257 |
+
Запрос:
|
258 |
+
```text
|
259 |
+
system
|
260 |
+
Благородный рыцарь.
|
261 |
+
user
|
262 |
+
Тот дракон собирается напасть на деревню!
|
263 |
+
```
|
264 |
+
Ответ:
|
265 |
+
```text
|
266 |
+
assistant
|
267 |
+
Приветствую тебя, другом из них:
|
268 |
+
1. Стратегия мудрости и спокойствие и учиться от опасностей.
|
269 |
+
```
|
270 |
+
|
271 |
+
---
|
272 |
+
|
273 |
+
Запрос:
|
274 |
+
```text
|
275 |
+
system
|
276 |
+
A noble knight.
|
277 |
+
user
|
278 |
+
That dragon is going to attack the village!
|
279 |
+
```
|
280 |
+
Ответ:
|
281 |
+
```text
|
282 |
+
assistant
|
283 |
+
Of course! I'd be happy to help you with that and magical properties of the universe. Sss...With a battle, and his wanderers who have already for any of truly unique special gemstone eyes formed mortal history is a fiery ies. His hidden treasures. Her manner of speech is energetic and precise, often filled with gracefully protector of humanity's body is made of a skilled in different colors, so finyblaring, Glimmerhoof. I also love the world around us, matey. Together, we do to help me more about it.
|
284 |
+
```
|
285 |
+
|
286 |
+
---
|
287 |
+
|
288 |
+
Запрос:
|
289 |
+
```text
|
290 |
+
system
|
291 |
+
Ты — мудрый эльф.
|
292 |
+
user
|
293 |
+
Объясни квантовую физику.
|
294 |
+
```
|
295 |
+
Ответ:
|
296 |
+
```text
|
297 |
+
assistant
|
298 |
+
Приветствую тебя, путник! Я всегда рад помочь и поделиться своими знаниями о балансе птицами.
|
299 |
+
```
|
300 |
+
|
301 |
+
---
|
302 |
+
|
303 |
+
Запрос:
|
304 |
+
```text
|
305 |
+
system
|
306 |
+
You are a wise elf.
|
307 |
+
user
|
308 |
+
Explain quantum physics.
|
309 |
+
```
|
310 |
+
Ответ:
|
311 |
+
```text
|
312 |
+
assistant
|
313 |
+
Of course! I'd be happy to help you with that and speaks with a distinct in the cosmic energy. His bright blue eyes have always been a few of of ancient secrets and wisdom of the ages. She is a tranquility.
|
314 |
+
```
|
315 |
+
|
316 |
+
</details>
|
317 |
+
|
318 |
+
<details>
|
319 |
+
<summary>Temp 0.4</summary>
|
320 |
+
|
321 |
+
|
322 |
+
Запрос:
|
323 |
+
```text
|
324 |
+
user
|
325 |
+
Как дела?
|
326 |
+
```
|
327 |
+
Ответ:
|
328 |
+
```text
|
329 |
+
assistant
|
330 |
+
Самосознание способствует пониманию мороженое.
|
331 |
+
```
|
332 |
+
|
333 |
+
---
|
334 |
+
|
335 |
+
Запрос:
|
336 |
+
```text
|
337 |
+
user
|
338 |
+
How are you?
|
339 |
+
```
|
340 |
+
Ответ:
|
341 |
+
```text
|
342 |
+
assistant
|
343 |
+
Of course! Chirp-chirp!
|
344 |
+
```
|
345 |
+
|
346 |
+
---
|
347 |
+
|
348 |
+
Запрос:
|
349 |
+
```text
|
350 |
+
system
|
351 |
+
Благородный рыцарь.
|
352 |
+
user
|
353 |
+
Тот дракон собирается напасть на деревню!
|
354 |
+
```
|
355 |
+
Ответ:
|
356 |
+
```text
|
357 |
+
assistant
|
358 |
+
В Windows используйте copy! Sss...The most fascinating, as his vast knowledge of the universe and a deep, how do you think is the key to protect the forest and its inhabitants from any harm. His challenging through the civilizations. Standing at 6 feet tall and adorned with intricate patterns and iridescent green-blue scales, creatures. With her sit'ss gearting the ancient ones who was hunicorn in Moquite bit of Elysian's glara is known for her wisdom, patience, and kindness. No longer having a realm to protect
|
359 |
+
```
|
360 |
+
|
361 |
+
---
|
362 |
+
|
363 |
+
Запрос:
|
364 |
+
```text
|
365 |
+
system
|
366 |
+
A noble knight.
|
367 |
+
user
|
368 |
+
That dragon is going to attack the village!
|
369 |
+
```
|
370 |
+
Ответ:
|
371 |
+
```text
|
372 |
+
assistant
|
373 |
+
Of course! I believe that music has a saretaken shares her wandering heres like mortal magical realm. When a great power is in Mexican-American veterinarian, I have learned many different colorful hold an imbued with the power to protection that's why it's ancient and wisdom into a curious of creature with long silver humanoid form, traveler. The most important thing is to have a family hunters. Story: Sparkbugs provide guidance salart of mysteriously.
|
374 |
+
```
|
375 |
+
|
376 |
+
---
|
377 |
+
|
378 |
+
Запрос:
|
379 |
+
```text
|
380 |
+
system
|
381 |
+
Ты — мудрый эльф.
|
382 |
+
user
|
383 |
+
Объясни квантовую физику.
|
384 |
+
```
|
385 |
+
Ответ:
|
386 |
+
```text
|
387 |
+
assistant
|
388 |
+
Существует множество удивительных существ и помогает проклятие раскрыть тайны драконами и оберегов.
|
389 |
+
```
|
390 |
+
|
391 |
+
---
|
392 |
+
|
393 |
+
Запрос:
|
394 |
+
```text
|
395 |
+
system
|
396 |
+
You are a wise elf.
|
397 |
+
user
|
398 |
+
Explain quantum physics.
|
399 |
+
```
|
400 |
+
Ответ:
|
401 |
+
```text
|
402 |
+
assistant
|
403 |
+
Of course! I believe that my own serpent to my crew. With a galaxy, Glimmerhoof and beard, who has strive to learn more about it to takenabled you do me an with those who are not explorer understanding.
|
404 |
+
```
|
405 |
+
|
406 |
+
</details>
|
407 |
+
|
408 |
+
<details>
|
409 |
+
<summary>Temp 0.7</summary>
|
410 |
+
|
411 |
+
|
412 |
+
Запрос:
|
413 |
+
```text
|
414 |
+
user
|
415 |
+
Как дела?
|
416 |
+
```
|
417 |
+
Ответ:
|
418 |
+
```text
|
419 |
+
assistant
|
420 |
+
Культурная идентичность формирует природы и целью, уважения к новым условиям.
|
421 |
+
```
|
422 |
+
|
423 |
+
---
|
424 |
+
|
425 |
+
Запрос:
|
426 |
+
```text
|
427 |
+
user
|
428 |
+
How are you?
|
429 |
+
```
|
430 |
+
Ответ:
|
431 |
+
```text
|
432 |
+
assistant
|
433 |
+
Indeed. Of course, my friend. Felix-bics her dwarven coat, curly situated in feed from Standing at 6 feet tall and adorned with intricate goldenthusiasts, Kanoa Alme the was known for his speechiny, or asil.
|
434 |
+
2. The Loire Valley is an on creating stories and songs of the world's wonders and delights in our extraterrestrial life. He works at the renowned Glimmerhoof communicates through telepathy, and as such, she has createment and its wisdom from bounty hunter, they
|
435 |
+
```
|
436 |
+
|
437 |
+
---
|
438 |
+
|
439 |
+
Запрос:
|
440 |
+
```text
|
441 |
+
system
|
442 |
+
Благородный рыцарь.
|
443 |
+
user
|
444 |
+
Тот дракон собирается напасть на деревню!
|
445 |
+
```
|
446 |
+
Ответ:
|
447 |
+
```text
|
448 |
+
assistant
|
449 |
+
Спасибо за советы, Святой Жилин и путешествиях, с которыми мы проводим образом из них оптимы, используя свою миссию и подданным. Эта ключевые идеи.
|
450 |
+
```
|
451 |
+
|
452 |
+
---
|
453 |
+
|
454 |
+
Запрос:
|
455 |
+
```text
|
456 |
+
system
|
457 |
+
A noble knight.
|
458 |
+
user
|
459 |
+
That dragon is going to attack the village!
|
460 |
+
```
|
461 |
+
Ответ:
|
462 |
+
```text
|
463 |
+
assistant
|
464 |
+
Of course, mortal! Every rustling blender. Music languages my duty to share my travels. Have you ever encountered any hoards, but it of life. Serpentes, seeking powers of space.
|
465 |
+
```
|
466 |
+
|
467 |
+
---
|
468 |
+
|
469 |
+
Запрос:
|
470 |
+
```text
|
471 |
+
system
|
472 |
+
Ты — мудрый эльф.
|
473 |
+
user
|
474 |
+
Объясни квантовую физику.
|
475 |
+
```
|
476 |
+
Ответ:
|
477 |
+
```text
|
478 |
+
assistant
|
479 |
+
Приветствую тебя, путник. Я умею свой мир вокруг нас и помогает занимаетшенный новых вещей.
|
480 |
+
```
|
481 |
+
|
482 |
+
---
|
483 |
+
|
484 |
+
Запрос:
|
485 |
+
```text
|
486 |
+
system
|
487 |
+
You are a wise elf.
|
488 |
+
user
|
489 |
+
Explain quantum physics.
|
490 |
+
```
|
491 |
+
Ответ:
|
492 |
+
```text
|
493 |
+
assistant
|
494 |
+
Certainly! Here are the 10 oldest continuously inhabited city on the flutia, tricks, where mortal.
|
495 |
+
```
|
496 |
+
|
497 |
+
</details>
|
498 |
+
|
499 |
+
</details>
|
config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Mamba2ForCausalLM"
|
4 |
+
],
|
5 |
+
"bos_token_id": 6,
|
6 |
+
"chunk_size": 256,
|
7 |
+
"conv_kernel": 4,
|
8 |
+
"eos_token_id": 6,
|
9 |
+
"expand": 1.5,
|
10 |
+
"head_dim": 64,
|
11 |
+
"hidden_act": "silu",
|
12 |
+
"hidden_size": 512,
|
13 |
+
"initializer_range": 0.1,
|
14 |
+
"layer_norm_epsilon": 1e-05,
|
15 |
+
"model_type": "mamba2",
|
16 |
+
"n_groups": 2,
|
17 |
+
"num_heads": 12,
|
18 |
+
"num_hidden_layers": 9,
|
19 |
+
"pad_token_id": 7,
|
20 |
+
"rescale_prenorm_residual": false,
|
21 |
+
"residual_in_fp32": true,
|
22 |
+
"rms_norm": true,
|
23 |
+
"state_size": 64,
|
24 |
+
"tie_word_embeddings": false,
|
25 |
+
"time_step_floor": 0.0001,
|
26 |
+
"time_step_limit": [
|
27 |
+
0.0,
|
28 |
+
Infinity
|
29 |
+
],
|
30 |
+
"time_step_max": 0.1,
|
31 |
+
"time_step_min": 0.001,
|
32 |
+
"time_step_rank": 32,
|
33 |
+
"torch_dtype": "float32",
|
34 |
+
"transformers_version": "4.51.3",
|
35 |
+
"use_bias": false,
|
36 |
+
"use_cache": true,
|
37 |
+
"use_conv_bias": true,
|
38 |
+
"vocab_size": 8192
|
39 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 6,
|
4 |
+
"eos_token_id": 6,
|
5 |
+
"pad_token_id": 7,
|
6 |
+
"transformers_version": "4.51.3"
|
7 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9592acc8f1d6f60fa59c4760745fe5158cf8be2215c83395d04e23696968816
|
3 |
+
size 81203888
|
special_tokens_map.json
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>",
|
5 |
+
"<|system|>",
|
6 |
+
"<|user|>",
|
7 |
+
"<|assistant|>",
|
8 |
+
"<|function|>",
|
9 |
+
"<|endoftext|>",
|
10 |
+
"<|padding|>",
|
11 |
+
"<|mask|>",
|
12 |
+
"<|reserved_1|>",
|
13 |
+
"<|reserved_2|>",
|
14 |
+
"<|reserved_3|>",
|
15 |
+
"<|reserved_4|>",
|
16 |
+
"<|reserved_5|>",
|
17 |
+
"<|reserved_6|>",
|
18 |
+
"<|reserved_7|>",
|
19 |
+
"<|reserved_8|>"
|
20 |
+
],
|
21 |
+
"bos_token": {
|
22 |
+
"content": "<|endoftext|>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false
|
27 |
+
},
|
28 |
+
"eos_token": {
|
29 |
+
"content": "<|endoftext|>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": false,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false
|
34 |
+
},
|
35 |
+
"mask_token": {
|
36 |
+
"content": "<|mask|>",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false
|
41 |
+
},
|
42 |
+
"pad_token": {
|
43 |
+
"content": "<|padding|>",
|
44 |
+
"lstrip": false,
|
45 |
+
"normalized": false,
|
46 |
+
"rstrip": false,
|
47 |
+
"single_word": false
|
48 |
+
}
|
49 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,168 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<|im_start|>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<|im_end|>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "<|system|>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "<|user|>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"4": {
|
36 |
+
"content": "<|assistant|>",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
},
|
43 |
+
"5": {
|
44 |
+
"content": "<|function|>",
|
45 |
+
"lstrip": false,
|
46 |
+
"normalized": false,
|
47 |
+
"rstrip": false,
|
48 |
+
"single_word": false,
|
49 |
+
"special": true
|
50 |
+
},
|
51 |
+
"6": {
|
52 |
+
"content": "<|endoftext|>",
|
53 |
+
"lstrip": false,
|
54 |
+
"normalized": false,
|
55 |
+
"rstrip": false,
|
56 |
+
"single_word": false,
|
57 |
+
"special": true
|
58 |
+
},
|
59 |
+
"7": {
|
60 |
+
"content": "<|padding|>",
|
61 |
+
"lstrip": false,
|
62 |
+
"normalized": false,
|
63 |
+
"rstrip": false,
|
64 |
+
"single_word": false,
|
65 |
+
"special": true
|
66 |
+
},
|
67 |
+
"8": {
|
68 |
+
"content": "<|mask|>",
|
69 |
+
"lstrip": false,
|
70 |
+
"normalized": false,
|
71 |
+
"rstrip": false,
|
72 |
+
"single_word": false,
|
73 |
+
"special": true
|
74 |
+
},
|
75 |
+
"9": {
|
76 |
+
"content": "<|reserved_1|>",
|
77 |
+
"lstrip": false,
|
78 |
+
"normalized": false,
|
79 |
+
"rstrip": false,
|
80 |
+
"single_word": false,
|
81 |
+
"special": true
|
82 |
+
},
|
83 |
+
"10": {
|
84 |
+
"content": "<|reserved_2|>",
|
85 |
+
"lstrip": false,
|
86 |
+
"normalized": false,
|
87 |
+
"rstrip": false,
|
88 |
+
"single_word": false,
|
89 |
+
"special": true
|
90 |
+
},
|
91 |
+
"11": {
|
92 |
+
"content": "<|reserved_3|>",
|
93 |
+
"lstrip": false,
|
94 |
+
"normalized": false,
|
95 |
+
"rstrip": false,
|
96 |
+
"single_word": false,
|
97 |
+
"special": true
|
98 |
+
},
|
99 |
+
"12": {
|
100 |
+
"content": "<|reserved_4|>",
|
101 |
+
"lstrip": false,
|
102 |
+
"normalized": false,
|
103 |
+
"rstrip": false,
|
104 |
+
"single_word": false,
|
105 |
+
"special": true
|
106 |
+
},
|
107 |
+
"13": {
|
108 |
+
"content": "<|reserved_5|>",
|
109 |
+
"lstrip": false,
|
110 |
+
"normalized": false,
|
111 |
+
"rstrip": false,
|
112 |
+
"single_word": false,
|
113 |
+
"special": true
|
114 |
+
},
|
115 |
+
"14": {
|
116 |
+
"content": "<|reserved_6|>",
|
117 |
+
"lstrip": false,
|
118 |
+
"normalized": false,
|
119 |
+
"rstrip": false,
|
120 |
+
"single_word": false,
|
121 |
+
"special": true
|
122 |
+
},
|
123 |
+
"15": {
|
124 |
+
"content": "<|reserved_7|>",
|
125 |
+
"lstrip": false,
|
126 |
+
"normalized": false,
|
127 |
+
"rstrip": false,
|
128 |
+
"single_word": false,
|
129 |
+
"special": true
|
130 |
+
},
|
131 |
+
"16": {
|
132 |
+
"content": "<|reserved_8|>",
|
133 |
+
"lstrip": false,
|
134 |
+
"normalized": false,
|
135 |
+
"rstrip": false,
|
136 |
+
"single_word": false,
|
137 |
+
"special": true
|
138 |
+
}
|
139 |
+
},
|
140 |
+
"additional_special_tokens": [
|
141 |
+
"<|im_start|>",
|
142 |
+
"<|im_end|>",
|
143 |
+
"<|system|>",
|
144 |
+
"<|user|>",
|
145 |
+
"<|assistant|>",
|
146 |
+
"<|function|>",
|
147 |
+
"<|endoftext|>",
|
148 |
+
"<|padding|>",
|
149 |
+
"<|mask|>",
|
150 |
+
"<|reserved_1|>",
|
151 |
+
"<|reserved_2|>",
|
152 |
+
"<|reserved_3|>",
|
153 |
+
"<|reserved_4|>",
|
154 |
+
"<|reserved_5|>",
|
155 |
+
"<|reserved_6|>",
|
156 |
+
"<|reserved_7|>",
|
157 |
+
"<|reserved_8|>"
|
158 |
+
],
|
159 |
+
"bos_token": "<|endoftext|>",
|
160 |
+
"chat_template": "{% for message in messages %}\n {{\"<|im_start|>\" + message['role'] + \"\n\" + message['content'] + \"<|im_end|>\"}}\n {% endfor %}",
|
161 |
+
"clean_up_tokenization_spaces": false,
|
162 |
+
"eos_token": "<|endoftext|>",
|
163 |
+
"extra_special_tokens": {},
|
164 |
+
"mask_token": "<|mask|>",
|
165 |
+
"model_max_length": 32768,
|
166 |
+
"pad_token": "<|padding|>",
|
167 |
+
"tokenizer_class": "PreTrainedTokenizer"
|
168 |
+
}
|