yspkm commited on
Commit
642171b
·
verified ·
1 Parent(s): b9843ac

Training completed!

Browse files
Files changed (4) hide show
  1. README.md +11 -2
  2. adapter_model.bin +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -11,10 +11,12 @@ model-index:
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/yspkm/PrunePath-LoRA/runs/t3jrvcg2)
15
  # Mistral-7B-Instruct-v0.3-lora-math
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
@@ -42,10 +44,17 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_steps: 100
45
- - num_epochs: 0.01
46
 
47
  ### Training results
48
 
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/yspkm/PrunePath-LoRA/runs/555sbh01)
15
  # Mistral-7B-Instruct-v0.3-lora-math
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.3461
20
 
21
  ## Model description
22
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_steps: 100
47
+ - num_epochs: 3
48
 
49
  ### Training results
50
 
51
+ | Training Loss | Epoch | Step | Validation Loss |
52
+ |:-------------:|:------:|:----:|:---------------:|
53
+ | 0.3882 | 0.5133 | 200 | 0.3837 |
54
+ | 0.3388 | 1.0266 | 400 | 0.3599 |
55
+ | 0.3189 | 1.5399 | 600 | 0.3498 |
56
+ | 0.2659 | 2.0533 | 800 | 0.3459 |
57
+ | 0.269 | 2.5666 | 1000 | 0.3461 |
58
 
59
 
60
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de5c4cde32578eaaaba50b824cd65b3359298be275058ac89d4f84192ed4160b
3
  size 335706186
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f46460780a99f4551c3454551cf4acc885492b431b7b100567980867c8d49a0
3
  size 335706186
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49cccafe22e2ca6c4a9763e0838ca6543db5f404591630020e04b3c33df1fb37
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa9d7bbf041d75bcd241ffba83d4a7d7ab8cfa8c3af4f5e40e420d4497696cb2
3
  size 335604696
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a90b61d1169226ddaac059e069593612af44f73a41524976206515b53a3f95a3
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5752f9e50e22eed46f894534628dc49f92828ae53289543fad74816f09991014
3
  size 5240