PEFT
Safetensors
English
qwen2
Generated from Trainer
Files changed (1) hide show
  1. README.md +95 -83
README.md CHANGED
@@ -1,84 +1,96 @@
1
- ---
2
- library_name: peft
3
- license: other
4
- base_model: Qwen/Qwen2.5-3B-Instruct
5
- tags:
6
- - generated_from_trainer
7
- model-index:
8
- - name: pancho-v1-qw25-3B-UNAMGS
9
- results: []
10
- datasets:
11
- - Magpie-Align/Magpie-Pro-MT-300K-v0.1
12
- - Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered
13
- language:
14
- - en
15
- ---
16
-
17
- # pancho-v1-qw25-3B-UNAMGS
18
-
19
- This model is a fine-tuned version of [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct):
20
- It achieves the following results on the evaluation set:
21
- - Loss: 0.6555
22
- ![pancho-v1-qw25-3B-UNAMGS](https://huggingface.co/fblgit/pancho-v1-qw25-3B-UNAMGS/resolve/main/pancho-v1-qw25-3B.png)
23
-
24
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
25
-
26
- ## Model description
27
- Trained with MagPie:
28
- - Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered
29
- - Magpie-Align/Magpie-Pro-MT-300K-v0.1
30
-
31
- UNA on MLPs `4, 10, 16, 22, 28`
32
-
33
- MGS on 3 Scales.
34
-
35
- Following https://arxiv.org/abs//2410.21228 facts.
36
-
37
- ## License & Derivatives
38
- Any derivative (sft, merges, etc) using **ANY** layer from this model **MUST** include either `UNA` or `MGS` or `PANCHO` in their model name in order to obtain a LICENSE for derivatives of this model.
39
-
40
- ### Training hyperparameters
41
-
42
- The following hyperparameters were used during training:
43
- - learning_rate: 2e-05
44
- - seed: 42
45
- - distributed_type: multi-GPU
46
- - num_devices: 8
47
- - total_train_batch_size: 256
48
- - total_eval_batch_size: 16
49
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
- - num_epochs: 1
51
-
52
- ### Training results
53
-
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:------:|:----:|:---------------:|
56
- | 1.2127 | 0.0015 | 1 | 0.8711 |
57
- | 0.9905 | 0.0509 | 35 | 0.7338 |
58
- | 0.9685 | 0.1019 | 70 | 0.7114 |
59
- | 0.9554 | 0.1528 | 105 | 0.6994 |
60
- | 0.9077 | 0.2037 | 140 | 0.6915 |
61
- | 0.9149 | 0.2547 | 175 | 0.6859 |
62
- | 0.9363 | 0.3056 | 210 | 0.6795 |
63
- | 0.8975 | 0.3566 | 245 | 0.6745 |
64
- | 0.9095 | 0.4075 | 280 | 0.6709 |
65
- | 0.9216 | 0.4584 | 315 | 0.6681 |
66
- | 0.9143 | 0.5094 | 350 | 0.6666 |
67
- | 0.8879 | 0.5603 | 385 | 0.6645 |
68
- | 0.9194 | 0.6112 | 420 | 0.6625 |
69
- | 0.9123 | 0.6622 | 455 | 0.6615 |
70
- | 0.9056 | 0.7131 | 490 | 0.6591 |
71
- | 0.9172 | 0.7641 | 525 | 0.6578 |
72
- | 0.886 | 0.8150 | 560 | 0.6566 |
73
- | 0.9155 | 0.8659 | 595 | 0.6568 |
74
- | 0.9029 | 0.9169 | 630 | 0.6560 |
75
- | 0.8942 | 0.9678 | 665 | 0.6555 |
76
-
77
-
78
- ### Framework versions
79
-
80
- - PEFT 0.13.2
81
- - Transformers 4.45.2
82
- - Pytorch 2.3.0+cu121
83
- - Datasets 3.0.1
 
 
 
 
 
 
 
 
 
 
 
 
84
  - Tokenizers 0.20.1#
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: Qwen/Qwen2.5-3B-Instruct
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - Magpie-Align/Magpie-Pro-MT-300K-v0.1
9
+ - Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered
10
+ language:
11
+ - zho
12
+ - eng
13
+ - fra
14
+ - spa
15
+ - por
16
+ - deu
17
+ - ita
18
+ - rus
19
+ - jpn
20
+ - kor
21
+ - vie
22
+ - tha
23
+ - ara
24
+ model-index:
25
+ - name: pancho-v1-qw25-3B-UNAMGS
26
+ results: []
27
+ ---
28
+
29
+ # pancho-v1-qw25-3B-UNAMGS
30
+
31
+ This model is a fine-tuned version of [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct):
32
+ It achieves the following results on the evaluation set:
33
+ - Loss: 0.6555
34
+ ![pancho-v1-qw25-3B-UNAMGS](https://huggingface.co/fblgit/pancho-v1-qw25-3B-UNAMGS/resolve/main/pancho-v1-qw25-3B.png)
35
+
36
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
37
+
38
+ ## Model description
39
+ Trained with MagPie:
40
+ - Magpie-Align/Magpie-Llama-3.1-Pro-MT-300K-Filtered
41
+ - Magpie-Align/Magpie-Pro-MT-300K-v0.1
42
+
43
+ UNA on MLPs `4, 10, 16, 22, 28`
44
+
45
+ MGS on 3 Scales.
46
+
47
+ Following https://arxiv.org/abs//2410.21228 facts.
48
+
49
+ ## License & Derivatives
50
+ Any derivative (sft, merges, etc) using **ANY** layer from this model **MUST** include either `UNA` or `MGS` or `PANCHO` in their model name in order to obtain a LICENSE for derivatives of this model.
51
+
52
+ ### Training hyperparameters
53
+
54
+ The following hyperparameters were used during training:
55
+ - learning_rate: 2e-05
56
+ - seed: 42
57
+ - distributed_type: multi-GPU
58
+ - num_devices: 8
59
+ - total_train_batch_size: 256
60
+ - total_eval_batch_size: 16
61
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
+ - num_epochs: 1
63
+
64
+ ### Training results
65
+
66
+ | Training Loss | Epoch | Step | Validation Loss |
67
+ |:-------------:|:------:|:----:|:---------------:|
68
+ | 1.2127 | 0.0015 | 1 | 0.8711 |
69
+ | 0.9905 | 0.0509 | 35 | 0.7338 |
70
+ | 0.9685 | 0.1019 | 70 | 0.7114 |
71
+ | 0.9554 | 0.1528 | 105 | 0.6994 |
72
+ | 0.9077 | 0.2037 | 140 | 0.6915 |
73
+ | 0.9149 | 0.2547 | 175 | 0.6859 |
74
+ | 0.9363 | 0.3056 | 210 | 0.6795 |
75
+ | 0.8975 | 0.3566 | 245 | 0.6745 |
76
+ | 0.9095 | 0.4075 | 280 | 0.6709 |
77
+ | 0.9216 | 0.4584 | 315 | 0.6681 |
78
+ | 0.9143 | 0.5094 | 350 | 0.6666 |
79
+ | 0.8879 | 0.5603 | 385 | 0.6645 |
80
+ | 0.9194 | 0.6112 | 420 | 0.6625 |
81
+ | 0.9123 | 0.6622 | 455 | 0.6615 |
82
+ | 0.9056 | 0.7131 | 490 | 0.6591 |
83
+ | 0.9172 | 0.7641 | 525 | 0.6578 |
84
+ | 0.886 | 0.8150 | 560 | 0.6566 |
85
+ | 0.9155 | 0.8659 | 595 | 0.6568 |
86
+ | 0.9029 | 0.9169 | 630 | 0.6560 |
87
+ | 0.8942 | 0.9678 | 665 | 0.6555 |
88
+
89
+
90
+ ### Framework versions
91
+
92
+ - PEFT 0.13.2
93
+ - Transformers 4.45.2
94
+ - Pytorch 2.3.0+cu121
95
+ - Datasets 3.0.1
96
  - Tokenizers 0.20.1#