Heralax commited on
Commit
750d114
·
verified ·
1 Parent(s): 7c75315

End of training

Browse files
Files changed (2) hide show
  1. README.md +193 -0
  2. generation_config.json +7 -0
README.md ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Heralax/demo-nursing-model-pretrain
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ datasets:
9
+ - axolotl_correction_conversations_hidden-openstax-nursing.json
10
+ - axolotl_rag_conversations_hidden-openstax-nursing.jsonl
11
+ - pretraining_subset_1421673.jsonl
12
+ - factual_sft_completion/combined_all_0.jsonl
13
+ - factual_sft_completion/combined_all_2.jsonl
14
+ - factual_sft_completion/combined_all_3.jsonl
15
+ - factual_sft_completion/combined_all_1.jsonl
16
+ - generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2012946.jsonl
17
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_503236.jsonl
18
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_251557.jsonl
19
+ - generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_1006231.jsonl
20
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_503236.jsonl
21
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_251557.jsonl
22
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_503115.jsonl
23
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_503236.jsonl
24
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_251557.jsonl
25
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1006473.jsonl
26
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_503236.jsonl
27
+ - generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_251557.jsonl
28
+ model-index:
29
+ - name: demo-nursing-model-sft-2
30
+ results: []
31
+ ---
32
+
33
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
34
+ should probably proofread and complete it, then remove this comment. -->
35
+
36
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
37
+ <details><summary>See axolotl config</summary>
38
+
39
+ axolotl version: `0.10.0.dev0`
40
+ ```yaml
41
+ base_model: Heralax/demo-nursing-model-pretrain
42
+ tokenizer_type: AutoTokenizer
43
+ model_type: AutoModelForCausalLM
44
+ load_in_8bit: false
45
+ load_in_4bit: false
46
+ strict: false
47
+ datasets:
48
+ - path: axolotl_correction_conversations_hidden-openstax-nursing.json
49
+ type: input_output
50
+ - path: axolotl_rag_conversations_hidden-openstax-nursing.jsonl
51
+ type: input_output
52
+ - path: pretraining_subset_1421673.jsonl
53
+ type: completion
54
+ - path: factual_sft_completion/combined_all_0.jsonl
55
+ type: completion
56
+ - path: factual_sft_completion/combined_all_2.jsonl
57
+ type: completion
58
+ - path: factual_sft_completion/combined_all_3.jsonl
59
+ type: completion
60
+ - path: factual_sft_completion/combined_all_1.jsonl
61
+ type: completion
62
+ - path: generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2012946.jsonl
63
+ type: completion
64
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_503236.jsonl
65
+ type: completion
66
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_251557.jsonl
67
+ type: completion
68
+ - path: generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_1006231.jsonl
69
+ type: completion
70
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_503236.jsonl
71
+ type: completion
72
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_251557.jsonl
73
+ type: completion
74
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_503115.jsonl
75
+ type: completion
76
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_503236.jsonl
77
+ type: completion
78
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_251557.jsonl
79
+ type: completion
80
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1006473.jsonl
81
+ type: completion
82
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_503236.jsonl
83
+ type: completion
84
+ - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_251557.jsonl
85
+ type: completion
86
+ dataset_prepared_path: last_finetune_prepared
87
+ output_dir: ./finetune-model-output
88
+ seed: 1337
89
+ sequence_len: 5000
90
+ sample_packing: true
91
+ pad_to_sequence_len: false
92
+ shuffle_merged_datasets: true
93
+ gradient_accumulation_steps: 75
94
+ micro_batch_size: 2
95
+ eval_batch_size: 4
96
+ num_epochs: 5
97
+ optimizer: paged_adamw_8bit
98
+ lr_scheduler: constant
99
+ learning_rate: 2.0e-05
100
+ noisy_embedding_alpha: 5
101
+ weight_decay: 0
102
+ train_on_inputs: false
103
+ group_by_length: false
104
+ bf16: true
105
+ fp16: false
106
+ tf32: false
107
+ gradient_checkpointing: true
108
+ logging_steps: 1
109
+ xformers_attention: false
110
+ flash_attention: true
111
+ chat_template: chatml
112
+ auto_resume_from_checkpoints: false
113
+ warmup_ratio: 0.1
114
+ evals_per_epoch: 1
115
+ val_set_size: 0.04
116
+ saves_per_epoch: 1
117
+ eval_sample_packing: false
118
+ save_total_limit: 2
119
+ special_tokens:
120
+ pad_token: <unk>
121
+ use_liger_kernel: true
122
+ plugins:
123
+ - axolotl.integrations.liger.LigerPlugin
124
+ liger_rope: true
125
+ liger_rms_norm: true
126
+ liger_glu_activation: true
127
+ liger_layer_norm: true
128
+ liger_fused_linear_cross_entropy: true
129
+ sequence_length: 10000
130
+ wandb_project: test-project
131
+ wandb_entity: ''
132
+ wandb_watch: ''
133
+ wandb_run_id: ''
134
+ wandb_log_model: ''
135
+ hub_model_id: Heralax/demo-nursing-model-sft-2
136
+ hub_strategy: all_checkpoints
137
+
138
+ ```
139
+
140
+ </details><br>
141
+
142
+ # demo-nursing-model-sft-2
143
+
144
+ This model is a fine-tuned version of [Heralax/demo-nursing-model-pretrain](https://huggingface.co/Heralax/demo-nursing-model-pretrain) on the axolotl_correction_conversations_hidden-openstax-nursing.json, the axolotl_rag_conversations_hidden-openstax-nursing.jsonl, the pretraining_subset_1421673.jsonl, the factual_sft_completion/combined_all_0.jsonl, the factual_sft_completion/combined_all_2.jsonl, the factual_sft_completion/combined_all_3.jsonl, the factual_sft_completion/combined_all_1.jsonl, the generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2012946.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_503236.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_251557.jsonl, the generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_1006231.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_503236.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_251557.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_503115.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_503236.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_251557.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1006473.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_503236.jsonl and the generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_251557.jsonl datasets.
145
+ It achieves the following results on the evaluation set:
146
+ - Loss: 0.8211
147
+
148
+ ## Model description
149
+
150
+ More information needed
151
+
152
+ ## Intended uses & limitations
153
+
154
+ More information needed
155
+
156
+ ## Training and evaluation data
157
+
158
+ More information needed
159
+
160
+ ## Training procedure
161
+
162
+ ### Training hyperparameters
163
+
164
+ The following hyperparameters were used during training:
165
+ - learning_rate: 2e-05
166
+ - train_batch_size: 2
167
+ - eval_batch_size: 4
168
+ - seed: 1337
169
+ - gradient_accumulation_steps: 75
170
+ - total_train_batch_size: 150
171
+ - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
172
+ - lr_scheduler_type: constant
173
+ - lr_scheduler_warmup_steps: 9
174
+ - training_steps: 90
175
+
176
+ ### Training results
177
+
178
+ | Training Loss | Epoch | Step | Validation Loss |
179
+ |:-------------:|:------:|:----:|:---------------:|
180
+ | 1.8069 | 0.0533 | 1 | 1.4362 |
181
+ | 0.7499 | 0.9602 | 18 | 0.6830 |
182
+ | 0.6717 | 1.9068 | 36 | 0.6260 |
183
+ | 0.3437 | 2.8535 | 54 | 0.6453 |
184
+ | 0.1937 | 3.8535 | 72 | 0.7476 |
185
+ | 0.2156 | 4.8535 | 90 | 0.8211 |
186
+
187
+
188
+ ### Framework versions
189
+
190
+ - Transformers 4.52.3
191
+ - Pytorch 2.6.0+cu124
192
+ - Datasets 3.6.0
193
+ - Tokenizers 0.21.1
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.52.3"
7
+ }