boda commited on
Commit
ab554da
·
verified ·
1 Parent(s): ab804c0

Training in progress, step 100

Browse files
adapter_config.json CHANGED
@@ -23,12 +23,12 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "o_proj",
27
  "q_proj",
28
- "up_proj",
29
  "k_proj",
 
30
  "v_proj",
31
- "gate_proj",
32
  "down_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "q_proj",
27
+ "gate_proj",
28
  "k_proj",
29
+ "o_proj",
30
  "v_proj",
31
+ "up_proj",
32
  "down_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a47a4671015ecaf760615c3f90857a849eadf11cf80466d56259a25f006b15b3
3
  size 80014016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7739a7b1af19d947453c720a75ed1b54f2add47dc33dde2e639090606104ada9
3
  size 80014016
tokenizer_config.json CHANGED
@@ -29,12 +29,12 @@
29
  }
30
  },
31
  "bos_token": "<s>",
32
- "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "extra_special_tokens": {},
36
  "legacy": true,
37
- "model_max_length": 2048,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
 
29
  }
30
  },
31
  "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "extra_special_tokens": {},
36
  "legacy": true,
37
+ "model_max_length": 4096,
38
  "pad_token": "</s>",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16a9d1d08ce057dd1b4d5448d3e0a681b09fc27b3bd31daeafc294534cdd140b
3
  size 7288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3f7502e160ef80e35db4010f4041fec247a56c4d888526b0c6e997e1a81f309
3
  size 7288