lianghsun commited on
Commit
ad1233d
·
1 Parent(s): 584fcde

Completed SFT training (5/5 epochs). Preparing for multi-round DPO training.

Browse files
README.md CHANGED
@@ -53,7 +53,7 @@ model-index:
53
  metrics:
54
  - name: single choice
55
  type: exact-match
56
- value: 28.71
57
  widget:
58
  - text: 中華民國憲法第一條
59
  metrics:
@@ -76,8 +76,9 @@ new_version: lianghsun/Llama-3.2-Taiwan-3B-Instruct
76
 
77
  | Update Date | Model Version | Key Changes |
78
  |--------------|-----------------------|-------------------------------------|
79
- | 2024/11/25 | v2024.11.25 | Updated model version to v2024.11.25, training progressed to (3/10) epochs. Still in SFT stage, DPO training remains pending. |
80
- | 2024/11/22 | v2024.11.22 |Initial upload: Model version v2024.11.22, training completed up to (1/10) epochs. Currently trained only on SFT, DPO training not yet performed. |
 
81
 
82
  </details>
83
 
@@ -223,17 +224,17 @@ docker run --runtime nvidia --gpus all \
223
  - **lr_scheduler_type:** cosine
224
  - **lr_scheduler_warmup_ratio:** 0.01
225
  - **num_epochs:** 5.0
 
226
 
227
  #### Speeds, Sizes, Times
228
 
229
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
230
- *(WIP)*
231
- - **Duration**:
232
- - **Train runtime**:
233
- - **Train samples per second**:
234
- - **Train steps per second**:
235
- - **Total training FLOPs**:
236
- - **Train loss**:
237
 
238
  ## Evaluation
239
 
@@ -347,4 +348,4 @@ base_model: lianghsun/Llama-3.2-Taiwan-3B-Instruct
347
  - Transformers 4.45.2
348
  - Pytorch 2.4.1+cu121
349
  - Datasets 2.21.0
350
- - Tokenizers 0.20.0
 
53
  metrics:
54
  - name: single choice
55
  type: exact-match
56
+ value: 31.1
57
  widget:
58
  - text: 中華民國憲法第一條
59
  metrics:
 
76
 
77
  | Update Date | Model Version | Key Changes |
78
  |--------------|-----------------------|-------------------------------------|
79
+ | 2024/11/27 | v2024.11.27 | Completed SFT training (5/5 epochs). Preparing for multi-round DPO training. |
80
+ | 2024/11/25 | v2024.11.25 | Updated model version to v2024.11.25, training progressed to (3/5) epochs. Still in SFT stage, DPO training remains pending. |
81
+ | 2024/11/22 | v2024.11.22 | Initial upload: Model version v2024.11.22, training completed up to (1/5) epochs. Currently trained only on SFT, DPO training not yet performed. |
82
 
83
  </details>
84
 
 
224
  - **lr_scheduler_type:** cosine
225
  - **lr_scheduler_warmup_ratio:** 0.01
226
  - **num_epochs:** 5.0
227
+ - **global_step:** 590
228
 
229
  #### Speeds, Sizes, Times
230
 
231
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
232
+ - **Duration**: 5 days, 16:15:11.17
233
+ - **Train runtime**: 490,511.1789
234
+ - **Train samples per second**: 25.37
235
+ - **Train steps per second**: 0.001
236
+ - **Total training FLOPs**: 26,658,386,120,540,160
237
+ - **Train loss**: 0.8533
 
238
 
239
  ## Evaluation
240
 
 
348
  - Transformers 4.45.2
349
  - Pytorch 2.4.1+cu121
350
  - Datasets 2.21.0
351
+ - Tokenizers 0.20.0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/owen/LLaMA-Factory/saves/llama3.2-3B/sft/full/sft_2024-11-21-1/checkpoint-10",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "lianghsun/Llama-3.2-Taiwan-3B-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:969c04d243dc2a80b40c3edb3e0391567710eecce407b22006c73fadbe8fb455
3
  size 4965799096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a4129d85cadd57a2bd201e67d424bf472aad2d1f708fc479d6888ff859dca4d
3
  size 4965799096
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c0133ef2cccfbf9d713da104bd8a27882674ad1bde198ae7e4b5b4d005e85df
3
  size 1459729952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f4a01c12a64842a84ab1aea11fc242f3cd7f305931c5e86385ca18896ede088
3
  size 1459729952