Heralax commited on
Commit
0d8324f
·
verified ·
1 Parent(s): d7cab4b

End of training

Browse files
Files changed (2) hide show
  1. README.md +169 -0
  2. generation_config.json +7 -0
README.md ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Heralax/test-model-4-pretrain
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ datasets:
9
+ - axolotl_rag_conversations_facts.jsonl
10
+ - axolotl_correction_conversations_facts.json
11
+ - pretraining_subset_2170418.jsonl
12
+ - factual_sft_completion/combined_all_0.jsonl
13
+ - factual_sft_completion/combined_all_1.jsonl
14
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl
15
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl
16
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl
17
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl
18
+ - generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl
19
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl
20
+ model-index:
21
+ - name: test-model-4-sft
22
+ results: []
23
+ ---
24
+
25
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26
+ should probably proofread and complete it, then remove this comment. -->
27
+
28
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
29
+ <details><summary>See axolotl config</summary>
30
+
31
+ axolotl version: `0.10.0.dev0`
32
+ ```yaml
33
+ base_model: Heralax/test-model-4-pretrain
34
+ tokenizer_type: AutoTokenizer
35
+ model_type: AutoModelForCausalLM
36
+ load_in_8bit: false
37
+ load_in_4bit: false
38
+ strict: false
39
+ datasets:
40
+ - path: axolotl_rag_conversations_facts.jsonl
41
+ type: input_output
42
+ - path: axolotl_correction_conversations_facts.json
43
+ type: input_output
44
+ - path: pretraining_subset_2170418.jsonl
45
+ type: completion
46
+ - path: factual_sft_completion/combined_all_0.jsonl
47
+ type: completion
48
+ - path: factual_sft_completion/combined_all_1.jsonl
49
+ type: completion
50
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl
51
+ type: completion
52
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl
53
+ type: completion
54
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl
55
+ type: completion
56
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl
57
+ type: completion
58
+ - path: generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl
59
+ type: completion
60
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl
61
+ type: completion
62
+ dataset_prepared_path: last_finetune_prepared
63
+ output_dir: ./finetune-model-output
64
+ seed: 1337
65
+ sequence_len: 5000
66
+ sample_packing: true
67
+ pad_to_sequence_len: false
68
+ shuffle_merged_datasets: true
69
+ gradient_accumulation_steps: 75
70
+ micro_batch_size: 2
71
+ eval_batch_size: 4
72
+ num_epochs: 5
73
+ optimizer: paged_adamw_8bit
74
+ lr_scheduler: constant
75
+ learning_rate: 2.0e-05
76
+ noisy_embedding_alpha: 5
77
+ weight_decay: 0
78
+ train_on_inputs: false
79
+ group_by_length: false
80
+ bf16: true
81
+ fp16: false
82
+ tf32: false
83
+ gradient_checkpointing: true
84
+ logging_steps: 1
85
+ xformers_attention: false
86
+ flash_attention: true
87
+ chat_template: chatml
88
+ auto_resume_from_checkpoints: false
89
+ warmup_ratio: 0.1
90
+ evals_per_epoch: 1
91
+ val_set_size: 0.04
92
+ saves_per_epoch: 1
93
+ eval_sample_packing: false
94
+ save_total_limit: 2
95
+ special_tokens:
96
+ pad_token: <unk>
97
+ use_liger_kernel: true
98
+ plugins:
99
+ - axolotl.integrations.liger.LigerPlugin
100
+ liger_rope: true
101
+ liger_rms_norm: true
102
+ liger_glu_activation: true
103
+ liger_layer_norm: true
104
+ liger_fused_linear_cross_entropy: true
105
+ sequence_length: 10000
106
+ wandb_project: test-project
107
+ wandb_entity: ''
108
+ wandb_watch: ''
109
+ wandb_run_id: ''
110
+ wandb_log_model: ''
111
+ hub_model_id: Heralax/test-model-4-sft
112
+ hub_strategy: all_checkpoints
113
+
114
+ ```
115
+
116
+ </details><br>
117
+
118
+ # test-model-4-sft
119
+
120
+ This model is a fine-tuned version of [Heralax/test-model-4-pretrain](https://huggingface.co/Heralax/test-model-4-pretrain) on the axolotl_rag_conversations_facts.jsonl, the axolotl_correction_conversations_facts.json, the pretraining_subset_2170418.jsonl, the factual_sft_completion/combined_all_0.jsonl, the factual_sft_completion/combined_all_1.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl and the generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl datasets.
121
+ It achieves the following results on the evaluation set:
122
+ - Loss: 0.6876
123
+
124
+ ## Model description
125
+
126
+ More information needed
127
+
128
+ ## Intended uses & limitations
129
+
130
+ More information needed
131
+
132
+ ## Training and evaluation data
133
+
134
+ More information needed
135
+
136
+ ## Training procedure
137
+
138
+ ### Training hyperparameters
139
+
140
+ The following hyperparameters were used during training:
141
+ - learning_rate: 2e-05
142
+ - train_batch_size: 2
143
+ - eval_batch_size: 4
144
+ - seed: 1337
145
+ - gradient_accumulation_steps: 75
146
+ - total_train_batch_size: 150
147
+ - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
148
+ - lr_scheduler_type: constant
149
+ - lr_scheduler_warmup_steps: 8
150
+ - training_steps: 85
151
+
152
+ ### Training results
153
+
154
+ | Training Loss | Epoch | Step | Validation Loss |
155
+ |:-------------:|:------:|:----:|:---------------:|
156
+ | 1.5402 | 0.0564 | 1 | 1.2586 |
157
+ | 0.5945 | 0.9594 | 17 | 0.5595 |
158
+ | 0.443 | 1.9029 | 34 | 0.5419 |
159
+ | 0.3117 | 2.8465 | 51 | 0.5845 |
160
+ | 0.1713 | 3.7901 | 68 | 0.6350 |
161
+ | 0.1231 | 4.7336 | 85 | 0.6876 |
162
+
163
+
164
+ ### Framework versions
165
+
166
+ - Transformers 4.52.3
167
+ - Pytorch 2.6.0+cu124
168
+ - Datasets 3.6.0
169
+ - Tokenizers 0.21.1
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.52.3"
7
+ }