li-muyang commited on
Commit
4de4786
·
verified ·
1 Parent(s): d75a9a5

Training in progress, step 100

Browse files
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "architectures": [
3
  "MistralForCausalLM"
4
  ],
@@ -20,7 +21,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.51.3",
24
  "use_cache": false,
25
  "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
21
  "sliding_window": 4096,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.45.2",
25
  "use_cache": false,
26
  "vocab_size": 32000
27
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f7e4b1377f4237728eedff5eafc5fa093e883f0a0ce628854023ad47250cb39
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ea83e4703f2c0aa29372ebd5e59008772e9d5e5328af4a54a616cd760f96bbc
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87b8f1f70fc27bfeed07af5d37035ce9795e3066191f40d3292fb4854c93bb31
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10123a28365033e1b65caff8176018580f6796c7f561b293f510206665279b9e
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2231d5759518371ba2dfd16a80d042c5434b129433a89aba8eafbed540605d6
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2470f7a7efccbda74e885de2861d8dbab3c9a663a81a2a3d1506a5d423caab03
3
  size 4540516344
tokenizer_config.json CHANGED
@@ -33,7 +33,6 @@
33
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
- "extra_special_tokens": {},
37
  "legacy": false,
38
  "model_max_length": 2048,
39
  "pad_token": "</s>",
 
33
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
 
36
  "legacy": false,
37
  "model_max_length": 2048,
38
  "pad_token": "</s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7716198cdd07965116a644e34c6d23c52e356a75c48e288a1efb5a9939f17052
3
- size 7096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db4d5c195da68f34521dab5074d51591c48492f2780514eeef9e830a23b91850
3
+ size 7032