root
commited on
Commit
·
dd68bae
1
Parent(s):
9204056
Initial commit
Browse files- config.json +129 -0
- generation_config.json +6 -0
- model-00001-of-00029.safetensors +3 -0
- model-00002-of-00029.safetensors +3 -0
- model-00003-of-00029.safetensors +3 -0
- model-00004-of-00029.safetensors +3 -0
- model-00005-of-00029.safetensors +3 -0
- model-00006-of-00029.safetensors +3 -0
- model-00007-of-00029.safetensors +3 -0
- model-00008-of-00029.safetensors +3 -0
- model-00009-of-00029.safetensors +3 -0
- model-00010-of-00029.safetensors +3 -0
- model-00011-of-00029.safetensors +3 -0
- model-00012-of-00029.safetensors +3 -0
- model-00013-of-00029.safetensors +3 -0
- model-00014-of-00029.safetensors +3 -0
- model-00015-of-00029.safetensors +3 -0
- model-00016-of-00029.safetensors +3 -0
- model-00017-of-00029.safetensors +3 -0
- model-00018-of-00029.safetensors +3 -0
- model-00019-of-00029.safetensors +3 -0
- model-00020-of-00029.safetensors +3 -0
- model-00021-of-00029.safetensors +3 -0
- model-00022-of-00029.safetensors +3 -0
- model-00023-of-00029.safetensors +3 -0
- model-00024-of-00029.safetensors +3 -0
- model-00025-of-00029.safetensors +3 -0
- model-00026-of-00029.safetensors +3 -0
- model-00027-of-00029.safetensors +3 -0
- model-00028-of-00029.safetensors +3 -0
- model-00029-of-00029.safetensors +3 -0
- model.safetensors.index.json +0 -0
- recipe.yaml +6 -0
- special_tokens_map.json +30 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +46 -0
config.json
ADDED
@@ -0,0 +1,129 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MixtralForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 1,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"head_dim": 128,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 6144,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 16384,
|
13 |
+
"max_position_embeddings": 65536,
|
14 |
+
"model_type": "mixtral",
|
15 |
+
"num_attention_heads": 48,
|
16 |
+
"num_experts_per_tok": 2,
|
17 |
+
"num_hidden_layers": 56,
|
18 |
+
"num_key_value_heads": 8,
|
19 |
+
"num_local_experts": 8,
|
20 |
+
"output_router_logits": false,
|
21 |
+
"quantization_config": {
|
22 |
+
"config_groups": {
|
23 |
+
"group_0": {
|
24 |
+
"input_activations": {
|
25 |
+
"actorder": null,
|
26 |
+
"block_structure": null,
|
27 |
+
"dynamic": false,
|
28 |
+
"group_size": null,
|
29 |
+
"num_bits": 8,
|
30 |
+
"observer": "minmax",
|
31 |
+
"observer_kwargs": {},
|
32 |
+
"strategy": "tensor",
|
33 |
+
"symmetric": true,
|
34 |
+
"type": "float"
|
35 |
+
},
|
36 |
+
"output_activations": null,
|
37 |
+
"targets": [
|
38 |
+
"Linear"
|
39 |
+
],
|
40 |
+
"weights": {
|
41 |
+
"actorder": null,
|
42 |
+
"block_structure": null,
|
43 |
+
"dynamic": false,
|
44 |
+
"group_size": null,
|
45 |
+
"num_bits": 8,
|
46 |
+
"observer": "minmax",
|
47 |
+
"observer_kwargs": {},
|
48 |
+
"strategy": "tensor",
|
49 |
+
"symmetric": true,
|
50 |
+
"type": "float"
|
51 |
+
}
|
52 |
+
}
|
53 |
+
},
|
54 |
+
"format": "float-quantized",
|
55 |
+
"global_compression_ratio": 1.558404548701109,
|
56 |
+
"ignore": [
|
57 |
+
"model.layers.0.block_sparse_moe.gate",
|
58 |
+
"model.layers.1.block_sparse_moe.gate",
|
59 |
+
"model.layers.2.block_sparse_moe.gate",
|
60 |
+
"model.layers.3.block_sparse_moe.gate",
|
61 |
+
"model.layers.4.block_sparse_moe.gate",
|
62 |
+
"model.layers.5.block_sparse_moe.gate",
|
63 |
+
"model.layers.6.block_sparse_moe.gate",
|
64 |
+
"model.layers.7.block_sparse_moe.gate",
|
65 |
+
"model.layers.8.block_sparse_moe.gate",
|
66 |
+
"model.layers.9.block_sparse_moe.gate",
|
67 |
+
"model.layers.10.block_sparse_moe.gate",
|
68 |
+
"model.layers.11.block_sparse_moe.gate",
|
69 |
+
"model.layers.12.block_sparse_moe.gate",
|
70 |
+
"model.layers.13.block_sparse_moe.gate",
|
71 |
+
"model.layers.14.block_sparse_moe.gate",
|
72 |
+
"model.layers.15.block_sparse_moe.gate",
|
73 |
+
"model.layers.16.block_sparse_moe.gate",
|
74 |
+
"model.layers.17.block_sparse_moe.gate",
|
75 |
+
"model.layers.18.block_sparse_moe.gate",
|
76 |
+
"model.layers.19.block_sparse_moe.gate",
|
77 |
+
"model.layers.20.block_sparse_moe.gate",
|
78 |
+
"model.layers.21.block_sparse_moe.gate",
|
79 |
+
"model.layers.22.block_sparse_moe.gate",
|
80 |
+
"model.layers.23.block_sparse_moe.gate",
|
81 |
+
"model.layers.24.block_sparse_moe.gate",
|
82 |
+
"model.layers.25.block_sparse_moe.gate",
|
83 |
+
"model.layers.26.block_sparse_moe.gate",
|
84 |
+
"model.layers.27.block_sparse_moe.gate",
|
85 |
+
"model.layers.28.block_sparse_moe.gate",
|
86 |
+
"model.layers.29.block_sparse_moe.gate",
|
87 |
+
"model.layers.30.block_sparse_moe.gate",
|
88 |
+
"model.layers.31.block_sparse_moe.gate",
|
89 |
+
"model.layers.32.block_sparse_moe.gate",
|
90 |
+
"model.layers.33.block_sparse_moe.gate",
|
91 |
+
"model.layers.34.block_sparse_moe.gate",
|
92 |
+
"model.layers.35.block_sparse_moe.gate",
|
93 |
+
"model.layers.36.block_sparse_moe.gate",
|
94 |
+
"model.layers.37.block_sparse_moe.gate",
|
95 |
+
"model.layers.38.block_sparse_moe.gate",
|
96 |
+
"model.layers.39.block_sparse_moe.gate",
|
97 |
+
"model.layers.40.block_sparse_moe.gate",
|
98 |
+
"model.layers.41.block_sparse_moe.gate",
|
99 |
+
"model.layers.42.block_sparse_moe.gate",
|
100 |
+
"model.layers.43.block_sparse_moe.gate",
|
101 |
+
"model.layers.44.block_sparse_moe.gate",
|
102 |
+
"model.layers.45.block_sparse_moe.gate",
|
103 |
+
"model.layers.46.block_sparse_moe.gate",
|
104 |
+
"model.layers.47.block_sparse_moe.gate",
|
105 |
+
"model.layers.48.block_sparse_moe.gate",
|
106 |
+
"model.layers.49.block_sparse_moe.gate",
|
107 |
+
"model.layers.50.block_sparse_moe.gate",
|
108 |
+
"model.layers.51.block_sparse_moe.gate",
|
109 |
+
"model.layers.52.block_sparse_moe.gate",
|
110 |
+
"model.layers.53.block_sparse_moe.gate",
|
111 |
+
"model.layers.54.block_sparse_moe.gate",
|
112 |
+
"model.layers.55.block_sparse_moe.gate",
|
113 |
+
"lm_head"
|
114 |
+
],
|
115 |
+
"kv_cache_scheme": null,
|
116 |
+
"quant_method": "compressed-tensors",
|
117 |
+
"quantization_status": "compressed"
|
118 |
+
},
|
119 |
+
"rms_norm_eps": 1e-05,
|
120 |
+
"rope_theta": 1000000,
|
121 |
+
"router_aux_loss_coef": 0.001,
|
122 |
+
"router_jitter_noise": 0.0,
|
123 |
+
"sliding_window": null,
|
124 |
+
"tie_word_embeddings": false,
|
125 |
+
"torch_dtype": "bfloat16",
|
126 |
+
"transformers_version": "4.50.0",
|
127 |
+
"use_cache": false,
|
128 |
+
"vocab_size": 32000
|
129 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.50.0"
|
6 |
+
}
|
model-00001-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a03bae71f4003ef63feea404e2c58cf07515426b7b20fe559ff0fc252804a19f
|
3 |
+
size 4998803472
|
model-00002-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d3b7a70fd57ffbe352c919536a297cf9af885f0233ec8337219c4b6f578319b
|
3 |
+
size 4907602348
|
model-00003-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02ec284dee2da83c2eb910ea36ba41ddfc142162adb0d6f85e4b9984044c818c
|
3 |
+
size 4907602332
|
model-00004-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25acf8cbbe61f96d7e0eef5c36d48af60133f0a53c41e68c77e21c60f67f2728
|
3 |
+
size 4907602316
|
model-00005-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:609cf30a5246a163a987eaec90c7e2a4c4f9a53bf00c8c6f71445528d04e3c56
|
3 |
+
size 4907602300
|
model-00006-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff010f78b8f264838c4047e2ec36d128da5e2bc436eeba3b4f44680dc962555c
|
3 |
+
size 4907602276
|
model-00007-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e3d2b4e80806d5b77ad9131356fdccbf4384e28cb4c8fc35fededd3fe098db0
|
3 |
+
size 4907602444
|
model-00008-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1872d8cc30ef61898ca5cc60100e079e0e7ff84565e818520ef9e45d771dcab4
|
3 |
+
size 4907602428
|
model-00009-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f44be62f97279515f933d0b0f3eadbe0235fe96c451ea2ebefd4684b3d7c549
|
3 |
+
size 4907602412
|
model-00010-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:444c2ccea8b2aea3aa76c8bf45d6eb0ef2a02ce6ebf3a8015225734165d83662
|
3 |
+
size 4907602396
|
model-00011-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:539a0354e8820e3080ff5344612d81ce1884d6d7ca5317735103149cee64aa0d
|
3 |
+
size 4907602380
|
model-00012-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29f233759a57923f0f8e1b8087e39aa5a25b7e9b9fb29f70b76b804e023b70ad
|
3 |
+
size 4907602364
|
model-00013-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e063bdd744110276ddc1126518630937de0cf220d0428f3a20a0da21a9d72be0
|
3 |
+
size 4907602348
|
model-00014-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c54c45aec5cf90b57ae812fbc69bc0ef35d3684557535d1bd56a060b573a3d40
|
3 |
+
size 4907602332
|
model-00015-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e7ba449e3af29aa39370d8f48aa1fcfbe361012a600123ace25972bd58bc577
|
3 |
+
size 4907602316
|
model-00016-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b0c7f7ea5eabd3c0106c44760347ae4b5c39042ec13eaafb97e3c2454fb2011
|
3 |
+
size 4907602300
|
model-00017-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3629742239796d962005d961ba5b508b5562c5e084d4c7840ceade9c68d0bf66
|
3 |
+
size 4907602284
|
model-00018-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f31361b272f9134cb09f45272f9af3336f49dce01d829c2070705de308c472f8
|
3 |
+
size 4907602268
|
model-00019-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f27dd8ae0707ca53bb620856b5bc84fb5bedba7c31decd9a93f33727059b5685
|
3 |
+
size 4907602252
|
model-00020-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d4946b5ad48491f9d586b88c0d4014b9337f6d71dad6ee726711229f0c359f0
|
3 |
+
size 4907602236
|
model-00021-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0426ded609b7c23035ac91ab4ffa7b9ea131daecbaf32541fda1cf3b848d0055
|
3 |
+
size 4907602220
|
model-00022-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9cb8a0a55e5db3a3a9495810a82b2643e79e20cf9b10663149384d983fe7eec
|
3 |
+
size 4970418388
|
model-00023-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd4df83081b0d4ae7fe3a120689637ceebb72e8f90abd724ca39a8dbd97b393e
|
3 |
+
size 4995682384
|
model-00024-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2f5c8376be690c808269e82fa133a35acedae4717d9b8afa3f180332d14f314
|
3 |
+
size 4970516900
|
model-00025-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:383f552959b325074ed36db432066f441b1f1bb2d33acdadd2a43ae173608be0
|
3 |
+
size 4907577508
|
model-00026-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:020244d20ba19f7862f17c63239e8b4862ea3a2e8023b47c25029eba91d20ad5
|
3 |
+
size 4907602572
|
model-00027-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:593e19d03edcd5aa168fba206b37cf191518216af6b31fe8f98351ee4a96b69f
|
3 |
+
size 4907602556
|
model-00028-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:786f80082a6d52b4db5c6643204c8d47ba6c89637cfa94c708ab49957b109e47
|
3 |
+
size 4907602540
|
model-00029-of-00029.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ceeabed3ab724f0929ecd9ad6b45d768d8729bd0dd02773731593de6aaf0cdad
|
3 |
+
size 3300041576
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
recipe.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DEFAULT_stage:
|
2 |
+
DEFAULT_modifiers:
|
3 |
+
QuantizationModifier:
|
4 |
+
ignore: [lm_head, 're:.*block_sparse_moe.gate']
|
5 |
+
targets: [Linear]
|
6 |
+
scheme: FP8
|
special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<unk>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
tokenizer_config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"add_prefix_space": null,
|
5 |
+
"added_tokens_decoder": {
|
6 |
+
"0": {
|
7 |
+
"content": "<unk>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false,
|
12 |
+
"special": true
|
13 |
+
},
|
14 |
+
"1": {
|
15 |
+
"content": "<s>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": false,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false,
|
20 |
+
"special": true
|
21 |
+
},
|
22 |
+
"2": {
|
23 |
+
"content": "</s>",
|
24 |
+
"lstrip": false,
|
25 |
+
"normalized": false,
|
26 |
+
"rstrip": false,
|
27 |
+
"single_word": false,
|
28 |
+
"special": true
|
29 |
+
}
|
30 |
+
},
|
31 |
+
"additional_special_tokens": [],
|
32 |
+
"bos_token": "<s>",
|
33 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{{ messages[0]['content'].strip() }}{% else %}{% set loop_messages = messages %}{{ 'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user\\'s questions.' }}{% endif %}{% for message in loop_messages %}{% if loop.index0 == 0 %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ ' USER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% else %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ '\nUSER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ ' ASSISTANT:' }}{% endif %}",
|
34 |
+
"clean_up_tokenization_spaces": false,
|
35 |
+
"eos_token": "</s>",
|
36 |
+
"extra_special_tokens": {},
|
37 |
+
"legacy": true,
|
38 |
+
"model_max_length": 1000000000000000019884624838656,
|
39 |
+
"pad_token": "<unk>",
|
40 |
+
"padding_side": "right",
|
41 |
+
"sp_model_kwargs": {},
|
42 |
+
"spaces_between_special_tokens": false,
|
43 |
+
"tokenizer_class": "LlamaTokenizer",
|
44 |
+
"unk_token": "<unk>",
|
45 |
+
"use_default_system_prompt": true
|
46 |
+
}
|