AlyGreo commited on
Commit
91c23f7
·
verified ·
1 Parent(s): 5f4b5f4

End of training

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/agreu77/huggingface/runs/d0or5tlg)
31
 
32
  This model was trained with SFT.
33
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/agreu77/huggingface/runs/64jcqo2e)
31
 
32
  This model was trained with SFT.
33
 
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77c2fcc90b63ee8c8888109c26331c21f4c1a749b096e00a46ce982a74ba5a19
3
- size 4994182536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abf278f991ceaf39bc671ca4c872694cf06b9cd59e67539ecad72247571bf27e
3
+ size 4957859248
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b14f6ea6e74e1d4fcf69875093d96d88b500bbcee1ea10770a217f5ad10fc9b
3
- size 240938000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f042708de529cc90e56bad5c6039ac86eb2e081943367a52626d15cad4474b2
3
+ size 283651216
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5235073536
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
@@ -261,7 +261,7 @@
261
  "model.layers.23.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00002.safetensors",
262
  "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
263
  "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
264
- "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
265
  "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
266
  "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
267
  "model.layers.24.post_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5241463296
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
 
261
  "model.layers.23.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00002.safetensors",
262
  "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
263
  "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
264
+ "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
265
  "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
266
  "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
267
  "model.layers.24.post_feedforward_layernorm.weight": "model-00002-of-00002.safetensors",
runs/Dec05_17-30-22_4184e4e168f8/events.out.tfevents.1733419842.4184e4e168f8.189.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fd8a8dccb476a2dae15176bf5597eb9c4fb55e1b0673a31bcd69bf182e0fe3a
3
+ size 5614
runs/Dec05_17-32-00_4184e4e168f8/events.out.tfevents.1733419939.4184e4e168f8.189.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bf1a186a3fb8092c1184802bc433d10e1e344d2e3a696dab7680696a7ee527f
3
+ size 7236
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8867d9ccf97412fda9e0d20d707c96c160f4d3b3cd034ee2bba92eb01cf84930
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0abe8d78075218fb2f0e35fe2ca31576d3241bac60637fd1c3274c4ee9dd486
3
  size 5560