li-muyang commited on
Commit
41c501f
·
verified ·
1 Parent(s): a0415e9

Training in progress, step 200

Browse files
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -21,7 +20,7 @@
21
  "sliding_window": 4096,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
- "transformers_version": "4.45.2",
25
- "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
1
  {
 
2
  "architectures": [
3
  "MistralForCausalLM"
4
  ],
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.51.3",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0dc444d6a681eacf8504708e69bac9805420cef1130e0231ebe0d979e2036b3
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb9e5c2b1f7ef335cd9c27a6d8e59446b10f78f947788cd9a0d69cb6e4a7fd3
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28e498aee42ca4b7813108d5f521d32d011f6835c4eccf1870b9949b228b2999
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16356b4d149127d756bf4b81b7d4220bbaf95619df18f0f1f3c08f0ad9e44fa7
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97cc05bfe47e9fcf6e75ff346a0767558df6eeae43bcd607801ae1c92e87268c
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd49ee5dcb9eb3b17be6658220f1d968f721dbe34f49c70e11e09a0b3b4a796f
3
  size 4540516344
runs/Apr15_22-22-29_nid002952/events.out.tfevents.1744727015.nid002952.616478.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d11dfef9c641978a22eed296a9a92940363bb9e1127aaec684a18cef55103cd
3
+ size 14330
tokenizer_config.json CHANGED
@@ -33,6 +33,7 @@
33
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
 
36
  "legacy": false,
37
  "model_max_length": 2048,
38
  "pad_token": "</s>",
 
33
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
+ "extra_special_tokens": {},
37
  "legacy": false,
38
  "model_max_length": 2048,
39
  "pad_token": "</s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26c6fda6f1dfe3aec161c08a25d2fa1064202063cc83a1312f83add3afefb0e4
3
- size 6968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7716198cdd07965116a644e34c6d23c52e356a75c48e288a1efb5a9939f17052
3
+ size 7096