nblinh63 commited on
Commit
331d223
·
verified ·
1 Parent(s): 74c1f8b

End of training

Browse files
README.md CHANGED
@@ -78,7 +78,7 @@ resume_from_checkpoint: null
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
- sequence_len: 2048
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
@@ -103,7 +103,7 @@ xformers_attention: true
103
 
104
  This model is a fine-tuned version of [migtissera/Tess-v2.5-Phi-3-medium-128k-14B](https://huggingface.co/migtissera/Tess-v2.5-Phi-3-medium-128k-14B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 0.9605
107
 
108
  ## Model description
109
 
@@ -135,7 +135,7 @@ The following hyperparameters were used during training:
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 0.6037 | 0.0001 | 10 | 0.9605 |
139
 
140
 
141
  ### Framework versions
 
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
+ sequence_len: 3072
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
 
103
 
104
  This model is a fine-tuned version of [migtissera/Tess-v2.5-Phi-3-medium-128k-14B](https://huggingface.co/migtissera/Tess-v2.5-Phi-3-medium-128k-14B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 0.9849
107
 
108
  ## Model description
109
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 0.6232 | 0.0001 | 10 | 0.9849 |
139
 
140
 
141
  ### Framework versions
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "gate_up_proj",
25
  "o_proj",
26
- "qkv_proj"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "o_proj",
24
+ "qkv_proj",
25
+ "down_proj",
26
+ "gate_up_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c22a8e3355c58e313a250838168aef40a6f2cd1d4993dccc693c90f9b44e027
3
  size 222938570
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:007ce58dbafc471d3589fc7cf59213bc4df296e525bf2b12b13418ed5669059e
3
  size 222938570
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:080266c775bcabbe53fe7ca8e5027836b5df61877e45ced7b9a233b0ff0f45d5
3
  size 222865880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fa3d21ebbdca5a713d5e1b070ed05604762d25f849cdda41bb04e8591a1aa13
3
  size 222865880
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c58eb57f3917b507923ea755e030c1beca7a855fad468c4aad6fe3913785ca85
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2c64a0499951c50351a09e07571b6e2d6b303197421e5759275754f342fd5b8
3
  size 6776