starsofchance commited on
Commit
a105bb3
·
verified ·
1 Parent(s): 04ae8cd

Uploaded lora adapters after finetuning on primevul

Browse files
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. ultival_mistral/checkpoint-4000/README.md +202 -0
  3. ultival_mistral/checkpoint-4000/adapter_config.json +39 -0
  4. ultival_mistral/checkpoint-4000/adapter_model.safetensors +3 -0
  5. ultival_mistral/checkpoint-4000/chat_template.jinja +1 -0
  6. ultival_mistral/checkpoint-4000/optimizer.pt +3 -0
  7. ultival_mistral/checkpoint-4000/rng_state.pth +3 -0
  8. ultival_mistral/checkpoint-4000/scheduler.pt +3 -0
  9. ultival_mistral/checkpoint-4000/special_tokens_map.json +24 -0
  10. ultival_mistral/checkpoint-4000/tokenizer.json +3 -0
  11. ultival_mistral/checkpoint-4000/tokenizer_config.json +0 -0
  12. ultival_mistral/checkpoint-4000/trainer_state.json +1466 -0
  13. ultival_mistral/checkpoint-4000/training_args.bin +3 -0
  14. ultival_mistral/checkpoint-5000/README.md +202 -0
  15. ultival_mistral/checkpoint-5000/adapter_config.json +39 -0
  16. ultival_mistral/checkpoint-5000/adapter_model.safetensors +3 -0
  17. ultival_mistral/checkpoint-5000/chat_template.jinja +1 -0
  18. ultival_mistral/checkpoint-5000/optimizer.pt +3 -0
  19. ultival_mistral/checkpoint-5000/rng_state.pth +3 -0
  20. ultival_mistral/checkpoint-5000/scheduler.pt +3 -0
  21. ultival_mistral/checkpoint-5000/special_tokens_map.json +24 -0
  22. ultival_mistral/checkpoint-5000/tokenizer.json +3 -0
  23. ultival_mistral/checkpoint-5000/tokenizer_config.json +0 -0
  24. ultival_mistral/checkpoint-5000/trainer_state.json +1824 -0
  25. ultival_mistral/checkpoint-5000/training_args.bin +3 -0
  26. ultival_mistral/checkpoint-6000/README.md +202 -0
  27. ultival_mistral/checkpoint-6000/adapter_config.json +39 -0
  28. ultival_mistral/checkpoint-6000/adapter_model.safetensors +3 -0
  29. ultival_mistral/checkpoint-6000/chat_template.jinja +1 -0
  30. ultival_mistral/checkpoint-6000/optimizer.pt +3 -0
  31. ultival_mistral/checkpoint-6000/rng_state.pth +3 -0
  32. ultival_mistral/checkpoint-6000/scheduler.pt +3 -0
  33. ultival_mistral/checkpoint-6000/special_tokens_map.json +24 -0
  34. ultival_mistral/checkpoint-6000/tokenizer.json +3 -0
  35. ultival_mistral/checkpoint-6000/tokenizer_config.json +0 -0
  36. ultival_mistral/checkpoint-6000/trainer_state.json +2182 -0
  37. ultival_mistral/checkpoint-6000/training_args.bin +3 -0
  38. ultival_mistral/checkpoint-6184/README.md +202 -0
  39. ultival_mistral/checkpoint-6184/adapter_config.json +39 -0
  40. ultival_mistral/checkpoint-6184/adapter_model.safetensors +3 -0
  41. ultival_mistral/checkpoint-6184/chat_template.jinja +1 -0
  42. ultival_mistral/checkpoint-6184/optimizer.pt +3 -0
  43. ultival_mistral/checkpoint-6184/rng_state.pth +3 -0
  44. ultival_mistral/checkpoint-6184/scheduler.pt +3 -0
  45. ultival_mistral/checkpoint-6184/special_tokens_map.json +24 -0
  46. ultival_mistral/checkpoint-6184/tokenizer.json +3 -0
  47. ultival_mistral/checkpoint-6184/tokenizer_config.json +0 -0
  48. ultival_mistral/checkpoint-6184/trainer_state.json +2245 -0
  49. ultival_mistral/checkpoint-6184/training_args.bin +3 -0
  50. ultival_mistral/runs/Jul03_16-12-32_a183f886ff78/events.out.tfevents.1751559215.a183f886ff78.1482.0 +3 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ ultival_mistral/checkpoint-4000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ ultival_mistral/checkpoint-5000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ ultival_mistral/checkpoint-6000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ ultival_mistral/checkpoint-6184/tokenizer.json filter=lfs diff=lfs merge=lfs -text
ultival_mistral/checkpoint-4000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Ministral-8B-Instruct-2410
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
ultival_mistral/checkpoint-4000/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Ministral-8B-Instruct-2410",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 32,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "o_proj",
28
+ "k_proj",
29
+ "down_proj",
30
+ "q_proj",
31
+ "gate_proj",
32
+ "up_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
ultival_mistral/checkpoint-4000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:698d9998ba9ac5614be57347cce32a6efd94fb9a4c4f85c88af56df8cc145037
3
+ size 349243752
ultival_mistral/checkpoint-4000/chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{% if messages[1]['role'] == 'user' %}{{ '[INST] ' + messages[0]['content'] + ' ' + messages[1]['content'] + ' [/INST]' }}{% set loop_messages = messages[2:] %}{% else %}{{ '[INST] ' + messages[0]['content'] + ' [/INST]' }}{% set loop_messages = messages[1:] %}{% endif %}{% else %}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}
ultival_mistral/checkpoint-4000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3db69e6e07f78ea590887f34eb2ad205b28add7aa3d0f2cb7d1b79888dc7be83
3
+ size 177919028
ultival_mistral/checkpoint-4000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd55c3b412ed63dda63d944e5148e12c72c1382a50017bf4d626548cc6ff19a6
3
+ size 14244
ultival_mistral/checkpoint-4000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ad2700215ffb525ca35a0aac5f8a10bb15a4b447f330bbaba9667f852d4d6d9
3
+ size 1064
ultival_mistral/checkpoint-4000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<pad>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
ultival_mistral/checkpoint-4000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be2965d3e435dbdef9376e5bdb3dcc2bfc7e271aa5152ae896cf53c7e920adf8
3
+ size 17078235
ultival_mistral/checkpoint-4000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
ultival_mistral/checkpoint-4000/trainer_state.json ADDED
@@ -0,0 +1,1466 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3000,
3
+ "best_metric": 0.5850913524627686,
4
+ "best_model_checkpoint": "./ultival_run2/checkpoint-3000",
5
+ "epoch": 1.2936610608020698,
6
+ "eval_steps": 1000,
7
+ "global_step": 4000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00646830530401035,
14
+ "grad_norm": 0.8976479172706604,
15
+ "learning_rate": 7.6e-06,
16
+ "loss": 1.5544,
17
+ "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.0129366106080207,
21
+ "grad_norm": 0.726784348487854,
22
+ "learning_rate": 1.56e-05,
23
+ "loss": 1.242,
24
+ "step": 40
25
+ },
26
+ {
27
+ "epoch": 0.019404915912031046,
28
+ "grad_norm": 0.4917074143886566,
29
+ "learning_rate": 2.36e-05,
30
+ "loss": 0.6694,
31
+ "step": 60
32
+ },
33
+ {
34
+ "epoch": 0.0258732212160414,
35
+ "grad_norm": 0.3471176326274872,
36
+ "learning_rate": 3.16e-05,
37
+ "loss": 0.5334,
38
+ "step": 80
39
+ },
40
+ {
41
+ "epoch": 0.03234152652005175,
42
+ "grad_norm": 0.41759616136550903,
43
+ "learning_rate": 3.960000000000001e-05,
44
+ "loss": 0.5435,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.03880983182406209,
49
+ "grad_norm": 0.3755040466785431,
50
+ "learning_rate": 4.76e-05,
51
+ "loss": 0.5684,
52
+ "step": 120
53
+ },
54
+ {
55
+ "epoch": 0.045278137128072445,
56
+ "grad_norm": 0.37058988213539124,
57
+ "learning_rate": 5.560000000000001e-05,
58
+ "loss": 0.5329,
59
+ "step": 140
60
+ },
61
+ {
62
+ "epoch": 0.0517464424320828,
63
+ "grad_norm": 0.42603132128715515,
64
+ "learning_rate": 6.36e-05,
65
+ "loss": 0.5457,
66
+ "step": 160
67
+ },
68
+ {
69
+ "epoch": 0.05821474773609314,
70
+ "grad_norm": 0.3154866099357605,
71
+ "learning_rate": 7.16e-05,
72
+ "loss": 0.5472,
73
+ "step": 180
74
+ },
75
+ {
76
+ "epoch": 0.0646830530401035,
77
+ "grad_norm": 0.3919835388660431,
78
+ "learning_rate": 7.960000000000001e-05,
79
+ "loss": 0.4675,
80
+ "step": 200
81
+ },
82
+ {
83
+ "epoch": 0.07115135834411385,
84
+ "grad_norm": 0.32180657982826233,
85
+ "learning_rate": 8.76e-05,
86
+ "loss": 0.5078,
87
+ "step": 220
88
+ },
89
+ {
90
+ "epoch": 0.07761966364812418,
91
+ "grad_norm": 0.3246982991695404,
92
+ "learning_rate": 9.56e-05,
93
+ "loss": 0.488,
94
+ "step": 240
95
+ },
96
+ {
97
+ "epoch": 0.08408796895213454,
98
+ "grad_norm": 0.341487318277359,
99
+ "learning_rate": 0.00010360000000000001,
100
+ "loss": 0.5058,
101
+ "step": 260
102
+ },
103
+ {
104
+ "epoch": 0.09055627425614489,
105
+ "grad_norm": 0.3597952127456665,
106
+ "learning_rate": 0.00011160000000000002,
107
+ "loss": 0.4574,
108
+ "step": 280
109
+ },
110
+ {
111
+ "epoch": 0.09702457956015524,
112
+ "grad_norm": 0.40912091732025146,
113
+ "learning_rate": 0.00011960000000000001,
114
+ "loss": 0.4931,
115
+ "step": 300
116
+ },
117
+ {
118
+ "epoch": 0.1034928848641656,
119
+ "grad_norm": 0.37430697679519653,
120
+ "learning_rate": 0.0001276,
121
+ "loss": 0.4724,
122
+ "step": 320
123
+ },
124
+ {
125
+ "epoch": 0.10996119016817593,
126
+ "grad_norm": 0.3920972943305969,
127
+ "learning_rate": 0.00013560000000000002,
128
+ "loss": 0.4997,
129
+ "step": 340
130
+ },
131
+ {
132
+ "epoch": 0.11642949547218628,
133
+ "grad_norm": 0.32511982321739197,
134
+ "learning_rate": 0.0001436,
135
+ "loss": 0.4889,
136
+ "step": 360
137
+ },
138
+ {
139
+ "epoch": 0.12289780077619664,
140
+ "grad_norm": 0.3926583528518677,
141
+ "learning_rate": 0.0001516,
142
+ "loss": 0.4772,
143
+ "step": 380
144
+ },
145
+ {
146
+ "epoch": 0.129366106080207,
147
+ "grad_norm": 0.3535909950733185,
148
+ "learning_rate": 0.0001596,
149
+ "loss": 0.465,
150
+ "step": 400
151
+ },
152
+ {
153
+ "epoch": 0.13583441138421734,
154
+ "grad_norm": 0.34381625056266785,
155
+ "learning_rate": 0.0001676,
156
+ "loss": 0.4934,
157
+ "step": 420
158
+ },
159
+ {
160
+ "epoch": 0.1423027166882277,
161
+ "grad_norm": 0.3753703236579895,
162
+ "learning_rate": 0.0001756,
163
+ "loss": 0.5068,
164
+ "step": 440
165
+ },
166
+ {
167
+ "epoch": 0.14877102199223805,
168
+ "grad_norm": 0.34895479679107666,
169
+ "learning_rate": 0.00018360000000000002,
170
+ "loss": 0.5186,
171
+ "step": 460
172
+ },
173
+ {
174
+ "epoch": 0.15523932729624837,
175
+ "grad_norm": 0.24720332026481628,
176
+ "learning_rate": 0.0001916,
177
+ "loss": 0.461,
178
+ "step": 480
179
+ },
180
+ {
181
+ "epoch": 0.16170763260025872,
182
+ "grad_norm": 0.3713696002960205,
183
+ "learning_rate": 0.0001996,
184
+ "loss": 0.4932,
185
+ "step": 500
186
+ },
187
+ {
188
+ "epoch": 0.16817593790426907,
189
+ "grad_norm": 0.3724994659423828,
190
+ "learning_rate": 0.00019933145672061927,
191
+ "loss": 0.5036,
192
+ "step": 520
193
+ },
194
+ {
195
+ "epoch": 0.17464424320827943,
196
+ "grad_norm": 0.3562956750392914,
197
+ "learning_rate": 0.00019862772695285012,
198
+ "loss": 0.4871,
199
+ "step": 540
200
+ },
201
+ {
202
+ "epoch": 0.18111254851228978,
203
+ "grad_norm": 0.331737756729126,
204
+ "learning_rate": 0.00019792399718508095,
205
+ "loss": 0.5178,
206
+ "step": 560
207
+ },
208
+ {
209
+ "epoch": 0.18758085381630013,
210
+ "grad_norm": 0.34049859642982483,
211
+ "learning_rate": 0.00019722026741731178,
212
+ "loss": 0.488,
213
+ "step": 580
214
+ },
215
+ {
216
+ "epoch": 0.19404915912031048,
217
+ "grad_norm": 0.36076661944389343,
218
+ "learning_rate": 0.00019651653764954258,
219
+ "loss": 0.4821,
220
+ "step": 600
221
+ },
222
+ {
223
+ "epoch": 0.20051746442432083,
224
+ "grad_norm": 0.3059399127960205,
225
+ "learning_rate": 0.0001958128078817734,
226
+ "loss": 0.4924,
227
+ "step": 620
228
+ },
229
+ {
230
+ "epoch": 0.2069857697283312,
231
+ "grad_norm": 0.30692920088768005,
232
+ "learning_rate": 0.00019510907811400423,
233
+ "loss": 0.4822,
234
+ "step": 640
235
+ },
236
+ {
237
+ "epoch": 0.21345407503234154,
238
+ "grad_norm": 0.3319448232650757,
239
+ "learning_rate": 0.00019440534834623506,
240
+ "loss": 0.4607,
241
+ "step": 660
242
+ },
243
+ {
244
+ "epoch": 0.21992238033635186,
245
+ "grad_norm": 0.3047769069671631,
246
+ "learning_rate": 0.00019370161857846588,
247
+ "loss": 0.5357,
248
+ "step": 680
249
+ },
250
+ {
251
+ "epoch": 0.22639068564036222,
252
+ "grad_norm": 0.4353258013725281,
253
+ "learning_rate": 0.00019299788881069668,
254
+ "loss": 0.4738,
255
+ "step": 700
256
+ },
257
+ {
258
+ "epoch": 0.23285899094437257,
259
+ "grad_norm": 0.30877336859703064,
260
+ "learning_rate": 0.00019229415904292754,
261
+ "loss": 0.5128,
262
+ "step": 720
263
+ },
264
+ {
265
+ "epoch": 0.23932729624838292,
266
+ "grad_norm": 0.3188530206680298,
267
+ "learning_rate": 0.00019159042927515836,
268
+ "loss": 0.4651,
269
+ "step": 740
270
+ },
271
+ {
272
+ "epoch": 0.24579560155239327,
273
+ "grad_norm": 0.40678778290748596,
274
+ "learning_rate": 0.0001908866995073892,
275
+ "loss": 0.5112,
276
+ "step": 760
277
+ },
278
+ {
279
+ "epoch": 0.2522639068564036,
280
+ "grad_norm": 0.34188327193260193,
281
+ "learning_rate": 0.00019018296973962,
282
+ "loss": 0.487,
283
+ "step": 780
284
+ },
285
+ {
286
+ "epoch": 0.258732212160414,
287
+ "grad_norm": 0.3002854585647583,
288
+ "learning_rate": 0.00018947923997185081,
289
+ "loss": 0.512,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.2652005174644243,
294
+ "grad_norm": 0.35756272077560425,
295
+ "learning_rate": 0.00018877551020408164,
296
+ "loss": 0.4812,
297
+ "step": 820
298
+ },
299
+ {
300
+ "epoch": 0.2716688227684347,
301
+ "grad_norm": 0.33527258038520813,
302
+ "learning_rate": 0.00018807178043631247,
303
+ "loss": 0.4895,
304
+ "step": 840
305
+ },
306
+ {
307
+ "epoch": 0.278137128072445,
308
+ "grad_norm": 0.3472883105278015,
309
+ "learning_rate": 0.0001873680506685433,
310
+ "loss": 0.4671,
311
+ "step": 860
312
+ },
313
+ {
314
+ "epoch": 0.2846054333764554,
315
+ "grad_norm": 0.3604671061038971,
316
+ "learning_rate": 0.0001866643209007741,
317
+ "loss": 0.5233,
318
+ "step": 880
319
+ },
320
+ {
321
+ "epoch": 0.2910737386804657,
322
+ "grad_norm": 0.31021153926849365,
323
+ "learning_rate": 0.00018596059113300492,
324
+ "loss": 0.491,
325
+ "step": 900
326
+ },
327
+ {
328
+ "epoch": 0.2975420439844761,
329
+ "grad_norm": 0.6417067050933838,
330
+ "learning_rate": 0.00018525686136523577,
331
+ "loss": 0.4858,
332
+ "step": 920
333
+ },
334
+ {
335
+ "epoch": 0.3040103492884864,
336
+ "grad_norm": 0.33719781041145325,
337
+ "learning_rate": 0.0001845531315974666,
338
+ "loss": 0.4722,
339
+ "step": 940
340
+ },
341
+ {
342
+ "epoch": 0.31047865459249674,
343
+ "grad_norm": 0.3568169176578522,
344
+ "learning_rate": 0.0001838494018296974,
345
+ "loss": 0.5088,
346
+ "step": 960
347
+ },
348
+ {
349
+ "epoch": 0.3169469598965071,
350
+ "grad_norm": 0.42162615060806274,
351
+ "learning_rate": 0.00018314567206192823,
352
+ "loss": 0.5091,
353
+ "step": 980
354
+ },
355
+ {
356
+ "epoch": 0.32341526520051744,
357
+ "grad_norm": 0.3212260603904724,
358
+ "learning_rate": 0.00018244194229415905,
359
+ "loss": 0.4654,
360
+ "step": 1000
361
+ },
362
+ {
363
+ "epoch": 0.32341526520051744,
364
+ "eval_loss": 0.6167479753494263,
365
+ "eval_runtime": 1013.3659,
366
+ "eval_samples_per_second": 6.046,
367
+ "eval_steps_per_second": 0.756,
368
+ "step": 1000
369
+ },
370
+ {
371
+ "epoch": 0.3298835705045278,
372
+ "grad_norm": 0.3184376060962677,
373
+ "learning_rate": 0.00018173821252638988,
374
+ "loss": 0.4712,
375
+ "step": 1020
376
+ },
377
+ {
378
+ "epoch": 0.33635187580853815,
379
+ "grad_norm": 0.3114645183086395,
380
+ "learning_rate": 0.0001810344827586207,
381
+ "loss": 0.4393,
382
+ "step": 1040
383
+ },
384
+ {
385
+ "epoch": 0.3428201811125485,
386
+ "grad_norm": 0.30122488737106323,
387
+ "learning_rate": 0.0001803307529908515,
388
+ "loss": 0.4882,
389
+ "step": 1060
390
+ },
391
+ {
392
+ "epoch": 0.34928848641655885,
393
+ "grad_norm": 0.2877393066883087,
394
+ "learning_rate": 0.00017962702322308233,
395
+ "loss": 0.413,
396
+ "step": 1080
397
+ },
398
+ {
399
+ "epoch": 0.35575679172056923,
400
+ "grad_norm": 0.33227887749671936,
401
+ "learning_rate": 0.00017892329345531319,
402
+ "loss": 0.4527,
403
+ "step": 1100
404
+ },
405
+ {
406
+ "epoch": 0.36222509702457956,
407
+ "grad_norm": 0.33423128724098206,
408
+ "learning_rate": 0.000178219563687544,
409
+ "loss": 0.4604,
410
+ "step": 1120
411
+ },
412
+ {
413
+ "epoch": 0.36869340232858994,
414
+ "grad_norm": 0.36922487616539,
415
+ "learning_rate": 0.0001775158339197748,
416
+ "loss": 0.5044,
417
+ "step": 1140
418
+ },
419
+ {
420
+ "epoch": 0.37516170763260026,
421
+ "grad_norm": 0.34816262125968933,
422
+ "learning_rate": 0.00017681210415200564,
423
+ "loss": 0.4898,
424
+ "step": 1160
425
+ },
426
+ {
427
+ "epoch": 0.3816300129366106,
428
+ "grad_norm": 0.32086044549942017,
429
+ "learning_rate": 0.00017610837438423646,
430
+ "loss": 0.4647,
431
+ "step": 1180
432
+ },
433
+ {
434
+ "epoch": 0.38809831824062097,
435
+ "grad_norm": 0.3879236876964569,
436
+ "learning_rate": 0.0001754046446164673,
437
+ "loss": 0.5179,
438
+ "step": 1200
439
+ },
440
+ {
441
+ "epoch": 0.3945666235446313,
442
+ "grad_norm": 0.2899722158908844,
443
+ "learning_rate": 0.00017470091484869812,
444
+ "loss": 0.4786,
445
+ "step": 1220
446
+ },
447
+ {
448
+ "epoch": 0.40103492884864167,
449
+ "grad_norm": 0.3844458758831024,
450
+ "learning_rate": 0.00017399718508092892,
451
+ "loss": 0.4513,
452
+ "step": 1240
453
+ },
454
+ {
455
+ "epoch": 0.407503234152652,
456
+ "grad_norm": 0.3678169548511505,
457
+ "learning_rate": 0.00017329345531315974,
458
+ "loss": 0.4789,
459
+ "step": 1260
460
+ },
461
+ {
462
+ "epoch": 0.4139715394566624,
463
+ "grad_norm": 0.37138116359710693,
464
+ "learning_rate": 0.00017258972554539057,
465
+ "loss": 0.4694,
466
+ "step": 1280
467
+ },
468
+ {
469
+ "epoch": 0.4204398447606727,
470
+ "grad_norm": 0.3604221045970917,
471
+ "learning_rate": 0.00017188599577762142,
472
+ "loss": 0.4496,
473
+ "step": 1300
474
+ },
475
+ {
476
+ "epoch": 0.4269081500646831,
477
+ "grad_norm": 0.33961811661720276,
478
+ "learning_rate": 0.00017118226600985222,
479
+ "loss": 0.4696,
480
+ "step": 1320
481
+ },
482
+ {
483
+ "epoch": 0.4333764553686934,
484
+ "grad_norm": 0.3244430124759674,
485
+ "learning_rate": 0.00017047853624208305,
486
+ "loss": 0.4322,
487
+ "step": 1340
488
+ },
489
+ {
490
+ "epoch": 0.4398447606727037,
491
+ "grad_norm": 0.2879267930984497,
492
+ "learning_rate": 0.00016977480647431388,
493
+ "loss": 0.4646,
494
+ "step": 1360
495
+ },
496
+ {
497
+ "epoch": 0.4463130659767141,
498
+ "grad_norm": 0.3437345027923584,
499
+ "learning_rate": 0.0001690710767065447,
500
+ "loss": 0.4711,
501
+ "step": 1380
502
+ },
503
+ {
504
+ "epoch": 0.45278137128072443,
505
+ "grad_norm": 0.3166882395744324,
506
+ "learning_rate": 0.00016836734693877553,
507
+ "loss": 0.486,
508
+ "step": 1400
509
+ },
510
+ {
511
+ "epoch": 0.4592496765847348,
512
+ "grad_norm": 0.3677555322647095,
513
+ "learning_rate": 0.00016766361717100633,
514
+ "loss": 0.4677,
515
+ "step": 1420
516
+ },
517
+ {
518
+ "epoch": 0.46571798188874514,
519
+ "grad_norm": 0.32657769322395325,
520
+ "learning_rate": 0.00016695988740323715,
521
+ "loss": 0.4649,
522
+ "step": 1440
523
+ },
524
+ {
525
+ "epoch": 0.4721862871927555,
526
+ "grad_norm": 0.2980731427669525,
527
+ "learning_rate": 0.00016625615763546798,
528
+ "loss": 0.4547,
529
+ "step": 1460
530
+ },
531
+ {
532
+ "epoch": 0.47865459249676584,
533
+ "grad_norm": 0.3690549433231354,
534
+ "learning_rate": 0.0001655524278676988,
535
+ "loss": 0.4512,
536
+ "step": 1480
537
+ },
538
+ {
539
+ "epoch": 0.4851228978007762,
540
+ "grad_norm": 0.31456947326660156,
541
+ "learning_rate": 0.00016484869809992963,
542
+ "loss": 0.4675,
543
+ "step": 1500
544
+ },
545
+ {
546
+ "epoch": 0.49159120310478654,
547
+ "grad_norm": 0.3823802173137665,
548
+ "learning_rate": 0.00016414496833216046,
549
+ "loss": 0.4849,
550
+ "step": 1520
551
+ },
552
+ {
553
+ "epoch": 0.49805950840879687,
554
+ "grad_norm": 0.33280324935913086,
555
+ "learning_rate": 0.0001634412385643913,
556
+ "loss": 0.4876,
557
+ "step": 1540
558
+ },
559
+ {
560
+ "epoch": 0.5045278137128072,
561
+ "grad_norm": 0.28788062930107117,
562
+ "learning_rate": 0.00016273750879662211,
563
+ "loss": 0.426,
564
+ "step": 1560
565
+ },
566
+ {
567
+ "epoch": 0.5109961190168176,
568
+ "grad_norm": 0.3809375762939453,
569
+ "learning_rate": 0.00016203377902885294,
570
+ "loss": 0.4676,
571
+ "step": 1580
572
+ },
573
+ {
574
+ "epoch": 0.517464424320828,
575
+ "grad_norm": 0.27940312027931213,
576
+ "learning_rate": 0.00016133004926108374,
577
+ "loss": 0.4413,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 0.5239327296248383,
582
+ "grad_norm": 0.30946919322013855,
583
+ "learning_rate": 0.00016062631949331457,
584
+ "loss": 0.4395,
585
+ "step": 1620
586
+ },
587
+ {
588
+ "epoch": 0.5304010349288486,
589
+ "grad_norm": 0.36931583285331726,
590
+ "learning_rate": 0.0001599225897255454,
591
+ "loss": 0.4565,
592
+ "step": 1640
593
+ },
594
+ {
595
+ "epoch": 0.536869340232859,
596
+ "grad_norm": 0.3652470111846924,
597
+ "learning_rate": 0.00015921885995777622,
598
+ "loss": 0.4522,
599
+ "step": 1660
600
+ },
601
+ {
602
+ "epoch": 0.5433376455368694,
603
+ "grad_norm": 0.3060782551765442,
604
+ "learning_rate": 0.00015851513019000705,
605
+ "loss": 0.45,
606
+ "step": 1680
607
+ },
608
+ {
609
+ "epoch": 0.5498059508408797,
610
+ "grad_norm": 0.3275803029537201,
611
+ "learning_rate": 0.00015781140042223787,
612
+ "loss": 0.445,
613
+ "step": 1700
614
+ },
615
+ {
616
+ "epoch": 0.55627425614489,
617
+ "grad_norm": 0.34865352511405945,
618
+ "learning_rate": 0.0001571076706544687,
619
+ "loss": 0.4306,
620
+ "step": 1720
621
+ },
622
+ {
623
+ "epoch": 0.5627425614489003,
624
+ "grad_norm": 0.42525383830070496,
625
+ "learning_rate": 0.00015640394088669953,
626
+ "loss": 0.4544,
627
+ "step": 1740
628
+ },
629
+ {
630
+ "epoch": 0.5692108667529108,
631
+ "grad_norm": 0.32165050506591797,
632
+ "learning_rate": 0.00015570021111893035,
633
+ "loss": 0.484,
634
+ "step": 1760
635
+ },
636
+ {
637
+ "epoch": 0.5756791720569211,
638
+ "grad_norm": 0.3213440477848053,
639
+ "learning_rate": 0.00015499648135116115,
640
+ "loss": 0.4469,
641
+ "step": 1780
642
+ },
643
+ {
644
+ "epoch": 0.5821474773609314,
645
+ "grad_norm": 0.42835432291030884,
646
+ "learning_rate": 0.00015429275158339198,
647
+ "loss": 0.4592,
648
+ "step": 1800
649
+ },
650
+ {
651
+ "epoch": 0.5886157826649417,
652
+ "grad_norm": 0.3861497938632965,
653
+ "learning_rate": 0.0001535890218156228,
654
+ "loss": 0.4607,
655
+ "step": 1820
656
+ },
657
+ {
658
+ "epoch": 0.5950840879689522,
659
+ "grad_norm": 0.4156148135662079,
660
+ "learning_rate": 0.00015288529204785363,
661
+ "loss": 0.4541,
662
+ "step": 1840
663
+ },
664
+ {
665
+ "epoch": 0.6015523932729625,
666
+ "grad_norm": 0.32371824979782104,
667
+ "learning_rate": 0.00015218156228008446,
668
+ "loss": 0.4287,
669
+ "step": 1860
670
+ },
671
+ {
672
+ "epoch": 0.6080206985769728,
673
+ "grad_norm": 0.363788366317749,
674
+ "learning_rate": 0.00015147783251231528,
675
+ "loss": 0.428,
676
+ "step": 1880
677
+ },
678
+ {
679
+ "epoch": 0.6144890038809832,
680
+ "grad_norm": 0.3932839334011078,
681
+ "learning_rate": 0.0001507741027445461,
682
+ "loss": 0.4359,
683
+ "step": 1900
684
+ },
685
+ {
686
+ "epoch": 0.6209573091849935,
687
+ "grad_norm": 0.3009307086467743,
688
+ "learning_rate": 0.00015007037297677694,
689
+ "loss": 0.3976,
690
+ "step": 1920
691
+ },
692
+ {
693
+ "epoch": 0.6274256144890039,
694
+ "grad_norm": 0.3484659194946289,
695
+ "learning_rate": 0.00014936664320900776,
696
+ "loss": 0.4584,
697
+ "step": 1940
698
+ },
699
+ {
700
+ "epoch": 0.6338939197930142,
701
+ "grad_norm": 0.33728504180908203,
702
+ "learning_rate": 0.00014866291344123856,
703
+ "loss": 0.4543,
704
+ "step": 1960
705
+ },
706
+ {
707
+ "epoch": 0.6403622250970246,
708
+ "grad_norm": 0.3207024037837982,
709
+ "learning_rate": 0.0001479591836734694,
710
+ "loss": 0.4611,
711
+ "step": 1980
712
+ },
713
+ {
714
+ "epoch": 0.6468305304010349,
715
+ "grad_norm": 0.29134076833724976,
716
+ "learning_rate": 0.00014725545390570022,
717
+ "loss": 0.439,
718
+ "step": 2000
719
+ },
720
+ {
721
+ "epoch": 0.6468305304010349,
722
+ "eval_loss": 0.6003134250640869,
723
+ "eval_runtime": 968.0302,
724
+ "eval_samples_per_second": 6.329,
725
+ "eval_steps_per_second": 0.791,
726
+ "step": 2000
727
+ },
728
+ {
729
+ "epoch": 0.6532988357050453,
730
+ "grad_norm": 0.3548874258995056,
731
+ "learning_rate": 0.00014655172413793104,
732
+ "loss": 0.4548,
733
+ "step": 2020
734
+ },
735
+ {
736
+ "epoch": 0.6597671410090556,
737
+ "grad_norm": 0.3707274794578552,
738
+ "learning_rate": 0.00014584799437016187,
739
+ "loss": 0.4271,
740
+ "step": 2040
741
+ },
742
+ {
743
+ "epoch": 0.666235446313066,
744
+ "grad_norm": 0.37172695994377136,
745
+ "learning_rate": 0.0001451442646023927,
746
+ "loss": 0.4626,
747
+ "step": 2060
748
+ },
749
+ {
750
+ "epoch": 0.6727037516170763,
751
+ "grad_norm": 0.3459303081035614,
752
+ "learning_rate": 0.00014444053483462352,
753
+ "loss": 0.4317,
754
+ "step": 2080
755
+ },
756
+ {
757
+ "epoch": 0.6791720569210866,
758
+ "grad_norm": 0.3477213680744171,
759
+ "learning_rate": 0.00014373680506685435,
760
+ "loss": 0.4243,
761
+ "step": 2100
762
+ },
763
+ {
764
+ "epoch": 0.685640362225097,
765
+ "grad_norm": 0.31847599148750305,
766
+ "learning_rate": 0.00014303307529908517,
767
+ "loss": 0.4234,
768
+ "step": 2120
769
+ },
770
+ {
771
+ "epoch": 0.6921086675291074,
772
+ "grad_norm": 0.3231470584869385,
773
+ "learning_rate": 0.00014232934553131597,
774
+ "loss": 0.4302,
775
+ "step": 2140
776
+ },
777
+ {
778
+ "epoch": 0.6985769728331177,
779
+ "grad_norm": 0.3745759427547455,
780
+ "learning_rate": 0.0001416256157635468,
781
+ "loss": 0.4533,
782
+ "step": 2160
783
+ },
784
+ {
785
+ "epoch": 0.705045278137128,
786
+ "grad_norm": 0.34834805130958557,
787
+ "learning_rate": 0.00014092188599577763,
788
+ "loss": 0.4748,
789
+ "step": 2180
790
+ },
791
+ {
792
+ "epoch": 0.7115135834411385,
793
+ "grad_norm": 0.3675827085971832,
794
+ "learning_rate": 0.00014021815622800845,
795
+ "loss": 0.4401,
796
+ "step": 2200
797
+ },
798
+ {
799
+ "epoch": 0.7179818887451488,
800
+ "grad_norm": 0.2839801013469696,
801
+ "learning_rate": 0.00013951442646023928,
802
+ "loss": 0.4332,
803
+ "step": 2220
804
+ },
805
+ {
806
+ "epoch": 0.7244501940491591,
807
+ "grad_norm": 0.27037495374679565,
808
+ "learning_rate": 0.00013881069669247008,
809
+ "loss": 0.4382,
810
+ "step": 2240
811
+ },
812
+ {
813
+ "epoch": 0.7309184993531694,
814
+ "grad_norm": 0.34917208552360535,
815
+ "learning_rate": 0.00013810696692470093,
816
+ "loss": 0.4251,
817
+ "step": 2260
818
+ },
819
+ {
820
+ "epoch": 0.7373868046571799,
821
+ "grad_norm": 0.39801281690597534,
822
+ "learning_rate": 0.00013740323715693176,
823
+ "loss": 0.4432,
824
+ "step": 2280
825
+ },
826
+ {
827
+ "epoch": 0.7438551099611902,
828
+ "grad_norm": 0.8932256102561951,
829
+ "learning_rate": 0.00013669950738916259,
830
+ "loss": 0.4577,
831
+ "step": 2300
832
+ },
833
+ {
834
+ "epoch": 0.7503234152652005,
835
+ "grad_norm": 0.35125109553337097,
836
+ "learning_rate": 0.00013599577762139339,
837
+ "loss": 0.4502,
838
+ "step": 2320
839
+ },
840
+ {
841
+ "epoch": 0.7567917205692108,
842
+ "grad_norm": 0.3012046813964844,
843
+ "learning_rate": 0.0001352920478536242,
844
+ "loss": 0.4513,
845
+ "step": 2340
846
+ },
847
+ {
848
+ "epoch": 0.7632600258732212,
849
+ "grad_norm": 0.3242063820362091,
850
+ "learning_rate": 0.00013458831808585504,
851
+ "loss": 0.4559,
852
+ "step": 2360
853
+ },
854
+ {
855
+ "epoch": 0.7697283311772316,
856
+ "grad_norm": 0.3523860275745392,
857
+ "learning_rate": 0.00013388458831808587,
858
+ "loss": 0.4434,
859
+ "step": 2380
860
+ },
861
+ {
862
+ "epoch": 0.7761966364812419,
863
+ "grad_norm": 0.29160013794898987,
864
+ "learning_rate": 0.0001331808585503167,
865
+ "loss": 0.44,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 0.7826649417852523,
870
+ "grad_norm": 0.3047971725463867,
871
+ "learning_rate": 0.0001324771287825475,
872
+ "loss": 0.4336,
873
+ "step": 2420
874
+ },
875
+ {
876
+ "epoch": 0.7891332470892626,
877
+ "grad_norm": 0.3192126750946045,
878
+ "learning_rate": 0.00013177339901477832,
879
+ "loss": 0.4495,
880
+ "step": 2440
881
+ },
882
+ {
883
+ "epoch": 0.795601552393273,
884
+ "grad_norm": 0.37997299432754517,
885
+ "learning_rate": 0.00013106966924700917,
886
+ "loss": 0.4685,
887
+ "step": 2460
888
+ },
889
+ {
890
+ "epoch": 0.8020698576972833,
891
+ "grad_norm": 0.2811839282512665,
892
+ "learning_rate": 0.00013036593947924,
893
+ "loss": 0.4206,
894
+ "step": 2480
895
+ },
896
+ {
897
+ "epoch": 0.8085381630012937,
898
+ "grad_norm": 0.3121369779109955,
899
+ "learning_rate": 0.0001296622097114708,
900
+ "loss": 0.425,
901
+ "step": 2500
902
+ },
903
+ {
904
+ "epoch": 0.815006468305304,
905
+ "grad_norm": 0.3134746253490448,
906
+ "learning_rate": 0.00012895847994370162,
907
+ "loss": 0.4498,
908
+ "step": 2520
909
+ },
910
+ {
911
+ "epoch": 0.8214747736093143,
912
+ "grad_norm": 0.36752375960350037,
913
+ "learning_rate": 0.00012825475017593245,
914
+ "loss": 0.4288,
915
+ "step": 2540
916
+ },
917
+ {
918
+ "epoch": 0.8279430789133247,
919
+ "grad_norm": 0.3915141224861145,
920
+ "learning_rate": 0.00012755102040816328,
921
+ "loss": 0.4605,
922
+ "step": 2560
923
+ },
924
+ {
925
+ "epoch": 0.8344113842173351,
926
+ "grad_norm": 0.3662431538105011,
927
+ "learning_rate": 0.00012684729064039408,
928
+ "loss": 0.4194,
929
+ "step": 2580
930
+ },
931
+ {
932
+ "epoch": 0.8408796895213454,
933
+ "grad_norm": 0.3387574851512909,
934
+ "learning_rate": 0.0001261435608726249,
935
+ "loss": 0.4258,
936
+ "step": 2600
937
+ },
938
+ {
939
+ "epoch": 0.8473479948253557,
940
+ "grad_norm": 0.38025060296058655,
941
+ "learning_rate": 0.00012543983110485573,
942
+ "loss": 0.4291,
943
+ "step": 2620
944
+ },
945
+ {
946
+ "epoch": 0.8538163001293662,
947
+ "grad_norm": 0.3517441749572754,
948
+ "learning_rate": 0.00012473610133708658,
949
+ "loss": 0.4384,
950
+ "step": 2640
951
+ },
952
+ {
953
+ "epoch": 0.8602846054333765,
954
+ "grad_norm": 0.33037152886390686,
955
+ "learning_rate": 0.0001240323715693174,
956
+ "loss": 0.4452,
957
+ "step": 2660
958
+ },
959
+ {
960
+ "epoch": 0.8667529107373868,
961
+ "grad_norm": 0.3645654618740082,
962
+ "learning_rate": 0.0001233286418015482,
963
+ "loss": 0.3992,
964
+ "step": 2680
965
+ },
966
+ {
967
+ "epoch": 0.8732212160413971,
968
+ "grad_norm": 0.32396814227104187,
969
+ "learning_rate": 0.00012262491203377904,
970
+ "loss": 0.4,
971
+ "step": 2700
972
+ },
973
+ {
974
+ "epoch": 0.8796895213454075,
975
+ "grad_norm": 0.3377726078033447,
976
+ "learning_rate": 0.00012192118226600986,
977
+ "loss": 0.4306,
978
+ "step": 2720
979
+ },
980
+ {
981
+ "epoch": 0.8861578266494179,
982
+ "grad_norm": 0.32345983386039734,
983
+ "learning_rate": 0.00012121745249824067,
984
+ "loss": 0.4225,
985
+ "step": 2740
986
+ },
987
+ {
988
+ "epoch": 0.8926261319534282,
989
+ "grad_norm": 0.34778884053230286,
990
+ "learning_rate": 0.0001205137227304715,
991
+ "loss": 0.4256,
992
+ "step": 2760
993
+ },
994
+ {
995
+ "epoch": 0.8990944372574385,
996
+ "grad_norm": 0.3861066997051239,
997
+ "learning_rate": 0.00011980999296270233,
998
+ "loss": 0.4465,
999
+ "step": 2780
1000
+ },
1001
+ {
1002
+ "epoch": 0.9055627425614489,
1003
+ "grad_norm": 0.3882784843444824,
1004
+ "learning_rate": 0.00011910626319493314,
1005
+ "loss": 0.4026,
1006
+ "step": 2800
1007
+ },
1008
+ {
1009
+ "epoch": 0.9120310478654593,
1010
+ "grad_norm": 0.4380407929420471,
1011
+ "learning_rate": 0.00011840253342716397,
1012
+ "loss": 0.4375,
1013
+ "step": 2820
1014
+ },
1015
+ {
1016
+ "epoch": 0.9184993531694696,
1017
+ "grad_norm": 0.35943394899368286,
1018
+ "learning_rate": 0.00011769880365939481,
1019
+ "loss": 0.4337,
1020
+ "step": 2840
1021
+ },
1022
+ {
1023
+ "epoch": 0.92496765847348,
1024
+ "grad_norm": 0.3862978219985962,
1025
+ "learning_rate": 0.00011699507389162563,
1026
+ "loss": 0.3949,
1027
+ "step": 2860
1028
+ },
1029
+ {
1030
+ "epoch": 0.9314359637774903,
1031
+ "grad_norm": 0.34592369198799133,
1032
+ "learning_rate": 0.00011629134412385645,
1033
+ "loss": 0.3798,
1034
+ "step": 2880
1035
+ },
1036
+ {
1037
+ "epoch": 0.9379042690815006,
1038
+ "grad_norm": 0.34942877292633057,
1039
+ "learning_rate": 0.00011558761435608727,
1040
+ "loss": 0.4235,
1041
+ "step": 2900
1042
+ },
1043
+ {
1044
+ "epoch": 0.944372574385511,
1045
+ "grad_norm": 0.28018903732299805,
1046
+ "learning_rate": 0.00011488388458831809,
1047
+ "loss": 0.4353,
1048
+ "step": 2920
1049
+ },
1050
+ {
1051
+ "epoch": 0.9508408796895214,
1052
+ "grad_norm": 0.3505672812461853,
1053
+ "learning_rate": 0.00011418015482054891,
1054
+ "loss": 0.3954,
1055
+ "step": 2940
1056
+ },
1057
+ {
1058
+ "epoch": 0.9573091849935317,
1059
+ "grad_norm": 0.3264220654964447,
1060
+ "learning_rate": 0.00011347642505277974,
1061
+ "loss": 0.4422,
1062
+ "step": 2960
1063
+ },
1064
+ {
1065
+ "epoch": 0.963777490297542,
1066
+ "grad_norm": 0.36382535099983215,
1067
+ "learning_rate": 0.00011277269528501055,
1068
+ "loss": 0.4116,
1069
+ "step": 2980
1070
+ },
1071
+ {
1072
+ "epoch": 0.9702457956015524,
1073
+ "grad_norm": 0.3051494359970093,
1074
+ "learning_rate": 0.00011206896551724138,
1075
+ "loss": 0.4135,
1076
+ "step": 3000
1077
+ },
1078
+ {
1079
+ "epoch": 0.9702457956015524,
1080
+ "eval_loss": 0.5850913524627686,
1081
+ "eval_runtime": 967.6256,
1082
+ "eval_samples_per_second": 6.332,
1083
+ "eval_steps_per_second": 0.792,
1084
+ "step": 3000
1085
+ },
1086
+ {
1087
+ "epoch": 0.9767141009055628,
1088
+ "grad_norm": 0.34533271193504333,
1089
+ "learning_rate": 0.00011136523574947222,
1090
+ "loss": 0.4312,
1091
+ "step": 3020
1092
+ },
1093
+ {
1094
+ "epoch": 0.9831824062095731,
1095
+ "grad_norm": 0.35221609473228455,
1096
+ "learning_rate": 0.00011066150598170305,
1097
+ "loss": 0.4245,
1098
+ "step": 3040
1099
+ },
1100
+ {
1101
+ "epoch": 0.9896507115135834,
1102
+ "grad_norm": 0.3893713057041168,
1103
+ "learning_rate": 0.00010995777621393386,
1104
+ "loss": 0.4048,
1105
+ "step": 3060
1106
+ },
1107
+ {
1108
+ "epoch": 0.9961190168175937,
1109
+ "grad_norm": 0.3586278557777405,
1110
+ "learning_rate": 0.00010925404644616468,
1111
+ "loss": 0.4159,
1112
+ "step": 3080
1113
+ },
1114
+ {
1115
+ "epoch": 1.0025873221216042,
1116
+ "grad_norm": 0.32620444893836975,
1117
+ "learning_rate": 0.0001085503166783955,
1118
+ "loss": 0.3741,
1119
+ "step": 3100
1120
+ },
1121
+ {
1122
+ "epoch": 1.0090556274256144,
1123
+ "grad_norm": 0.33812645077705383,
1124
+ "learning_rate": 0.00010784658691062632,
1125
+ "loss": 0.3694,
1126
+ "step": 3120
1127
+ },
1128
+ {
1129
+ "epoch": 1.0155239327296248,
1130
+ "grad_norm": 0.39038708806037903,
1131
+ "learning_rate": 0.00010714285714285715,
1132
+ "loss": 0.3882,
1133
+ "step": 3140
1134
+ },
1135
+ {
1136
+ "epoch": 1.0219922380336353,
1137
+ "grad_norm": 0.3209969997406006,
1138
+ "learning_rate": 0.00010643912737508796,
1139
+ "loss": 0.3437,
1140
+ "step": 3160
1141
+ },
1142
+ {
1143
+ "epoch": 1.0284605433376455,
1144
+ "grad_norm": 0.3705078661441803,
1145
+ "learning_rate": 0.00010573539760731879,
1146
+ "loss": 0.3568,
1147
+ "step": 3180
1148
+ },
1149
+ {
1150
+ "epoch": 1.034928848641656,
1151
+ "grad_norm": 0.3562203645706177,
1152
+ "learning_rate": 0.0001050316678395496,
1153
+ "loss": 0.3916,
1154
+ "step": 3200
1155
+ },
1156
+ {
1157
+ "epoch": 1.0413971539456663,
1158
+ "grad_norm": 0.3724074363708496,
1159
+ "learning_rate": 0.00010432793807178046,
1160
+ "loss": 0.3866,
1161
+ "step": 3220
1162
+ },
1163
+ {
1164
+ "epoch": 1.0478654592496766,
1165
+ "grad_norm": 0.40453559160232544,
1166
+ "learning_rate": 0.00010362420830401127,
1167
+ "loss": 0.368,
1168
+ "step": 3240
1169
+ },
1170
+ {
1171
+ "epoch": 1.054333764553687,
1172
+ "grad_norm": 0.3316144049167633,
1173
+ "learning_rate": 0.0001029204785362421,
1174
+ "loss": 0.3696,
1175
+ "step": 3260
1176
+ },
1177
+ {
1178
+ "epoch": 1.0608020698576972,
1179
+ "grad_norm": 0.3503650724887848,
1180
+ "learning_rate": 0.00010221674876847291,
1181
+ "loss": 0.3644,
1182
+ "step": 3280
1183
+ },
1184
+ {
1185
+ "epoch": 1.0672703751617076,
1186
+ "grad_norm": 0.3136923909187317,
1187
+ "learning_rate": 0.00010151301900070374,
1188
+ "loss": 0.3414,
1189
+ "step": 3300
1190
+ },
1191
+ {
1192
+ "epoch": 1.073738680465718,
1193
+ "grad_norm": 0.3573864996433258,
1194
+ "learning_rate": 0.00010080928923293456,
1195
+ "loss": 0.3586,
1196
+ "step": 3320
1197
+ },
1198
+ {
1199
+ "epoch": 1.0802069857697283,
1200
+ "grad_norm": 0.26215964555740356,
1201
+ "learning_rate": 0.00010010555946516538,
1202
+ "loss": 0.356,
1203
+ "step": 3340
1204
+ },
1205
+ {
1206
+ "epoch": 1.0866752910737387,
1207
+ "grad_norm": 0.3103466331958771,
1208
+ "learning_rate": 9.940182969739622e-05,
1209
+ "loss": 0.3478,
1210
+ "step": 3360
1211
+ },
1212
+ {
1213
+ "epoch": 1.093143596377749,
1214
+ "grad_norm": 0.3601842224597931,
1215
+ "learning_rate": 9.869809992962703e-05,
1216
+ "loss": 0.3635,
1217
+ "step": 3380
1218
+ },
1219
+ {
1220
+ "epoch": 1.0996119016817594,
1221
+ "grad_norm": 0.3715918958187103,
1222
+ "learning_rate": 9.799437016185785e-05,
1223
+ "loss": 0.3489,
1224
+ "step": 3400
1225
+ },
1226
+ {
1227
+ "epoch": 1.1060802069857698,
1228
+ "grad_norm": 0.4216213524341583,
1229
+ "learning_rate": 9.729064039408867e-05,
1230
+ "loss": 0.3461,
1231
+ "step": 3420
1232
+ },
1233
+ {
1234
+ "epoch": 1.11254851228978,
1235
+ "grad_norm": 0.3878704905509949,
1236
+ "learning_rate": 9.65869106263195e-05,
1237
+ "loss": 0.3548,
1238
+ "step": 3440
1239
+ },
1240
+ {
1241
+ "epoch": 1.1190168175937905,
1242
+ "grad_norm": 0.36658766865730286,
1243
+ "learning_rate": 9.588318085855032e-05,
1244
+ "loss": 0.3395,
1245
+ "step": 3460
1246
+ },
1247
+ {
1248
+ "epoch": 1.1254851228978007,
1249
+ "grad_norm": 0.38134855031967163,
1250
+ "learning_rate": 9.517945109078115e-05,
1251
+ "loss": 0.3829,
1252
+ "step": 3480
1253
+ },
1254
+ {
1255
+ "epoch": 1.131953428201811,
1256
+ "grad_norm": 0.42683905363082886,
1257
+ "learning_rate": 9.447572132301197e-05,
1258
+ "loss": 0.3914,
1259
+ "step": 3500
1260
+ },
1261
+ {
1262
+ "epoch": 1.1384217335058215,
1263
+ "grad_norm": 0.3340086042881012,
1264
+ "learning_rate": 9.377199155524279e-05,
1265
+ "loss": 0.3301,
1266
+ "step": 3520
1267
+ },
1268
+ {
1269
+ "epoch": 1.1448900388098318,
1270
+ "grad_norm": 0.37897104024887085,
1271
+ "learning_rate": 9.306826178747361e-05,
1272
+ "loss": 0.3536,
1273
+ "step": 3540
1274
+ },
1275
+ {
1276
+ "epoch": 1.1513583441138422,
1277
+ "grad_norm": 0.361061692237854,
1278
+ "learning_rate": 9.236453201970444e-05,
1279
+ "loss": 0.338,
1280
+ "step": 3560
1281
+ },
1282
+ {
1283
+ "epoch": 1.1578266494178524,
1284
+ "grad_norm": 0.34730997681617737,
1285
+ "learning_rate": 9.166080225193527e-05,
1286
+ "loss": 0.3679,
1287
+ "step": 3580
1288
+ },
1289
+ {
1290
+ "epoch": 1.1642949547218628,
1291
+ "grad_norm": 0.29455122351646423,
1292
+ "learning_rate": 9.095707248416608e-05,
1293
+ "loss": 0.3602,
1294
+ "step": 3600
1295
+ },
1296
+ {
1297
+ "epoch": 1.1707632600258733,
1298
+ "grad_norm": 0.33159056305885315,
1299
+ "learning_rate": 9.02533427163969e-05,
1300
+ "loss": 0.3359,
1301
+ "step": 3620
1302
+ },
1303
+ {
1304
+ "epoch": 1.1772315653298835,
1305
+ "grad_norm": 0.48726096749305725,
1306
+ "learning_rate": 8.954961294862773e-05,
1307
+ "loss": 0.3879,
1308
+ "step": 3640
1309
+ },
1310
+ {
1311
+ "epoch": 1.183699870633894,
1312
+ "grad_norm": 0.3182481825351715,
1313
+ "learning_rate": 8.884588318085856e-05,
1314
+ "loss": 0.3392,
1315
+ "step": 3660
1316
+ },
1317
+ {
1318
+ "epoch": 1.1901681759379044,
1319
+ "grad_norm": 0.3488766551017761,
1320
+ "learning_rate": 8.814215341308939e-05,
1321
+ "loss": 0.3419,
1322
+ "step": 3680
1323
+ },
1324
+ {
1325
+ "epoch": 1.1966364812419146,
1326
+ "grad_norm": 0.34345540404319763,
1327
+ "learning_rate": 8.74384236453202e-05,
1328
+ "loss": 0.3598,
1329
+ "step": 3700
1330
+ },
1331
+ {
1332
+ "epoch": 1.203104786545925,
1333
+ "grad_norm": 0.37642374634742737,
1334
+ "learning_rate": 8.673469387755102e-05,
1335
+ "loss": 0.3578,
1336
+ "step": 3720
1337
+ },
1338
+ {
1339
+ "epoch": 1.2095730918499352,
1340
+ "grad_norm": 0.3726005554199219,
1341
+ "learning_rate": 8.603096410978185e-05,
1342
+ "loss": 0.3547,
1343
+ "step": 3740
1344
+ },
1345
+ {
1346
+ "epoch": 1.2160413971539457,
1347
+ "grad_norm": 0.36603832244873047,
1348
+ "learning_rate": 8.532723434201268e-05,
1349
+ "loss": 0.3831,
1350
+ "step": 3760
1351
+ },
1352
+ {
1353
+ "epoch": 1.222509702457956,
1354
+ "grad_norm": 0.345103919506073,
1355
+ "learning_rate": 8.462350457424349e-05,
1356
+ "loss": 0.3563,
1357
+ "step": 3780
1358
+ },
1359
+ {
1360
+ "epoch": 1.2289780077619663,
1361
+ "grad_norm": 0.31071504950523376,
1362
+ "learning_rate": 8.391977480647432e-05,
1363
+ "loss": 0.3263,
1364
+ "step": 3800
1365
+ },
1366
+ {
1367
+ "epoch": 1.2354463130659767,
1368
+ "grad_norm": 0.3534263074398041,
1369
+ "learning_rate": 8.321604503870513e-05,
1370
+ "loss": 0.3591,
1371
+ "step": 3820
1372
+ },
1373
+ {
1374
+ "epoch": 1.2419146183699872,
1375
+ "grad_norm": 0.3568720817565918,
1376
+ "learning_rate": 8.251231527093597e-05,
1377
+ "loss": 0.3815,
1378
+ "step": 3840
1379
+ },
1380
+ {
1381
+ "epoch": 1.2483829236739974,
1382
+ "grad_norm": 0.3450963795185089,
1383
+ "learning_rate": 8.18085855031668e-05,
1384
+ "loss": 0.3481,
1385
+ "step": 3860
1386
+ },
1387
+ {
1388
+ "epoch": 1.2548512289780078,
1389
+ "grad_norm": 0.3554648160934448,
1390
+ "learning_rate": 8.110485573539761e-05,
1391
+ "loss": 0.3364,
1392
+ "step": 3880
1393
+ },
1394
+ {
1395
+ "epoch": 1.261319534282018,
1396
+ "grad_norm": 0.4188801348209381,
1397
+ "learning_rate": 8.040112596762844e-05,
1398
+ "loss": 0.3686,
1399
+ "step": 3900
1400
+ },
1401
+ {
1402
+ "epoch": 1.2677878395860285,
1403
+ "grad_norm": 0.3129253089427948,
1404
+ "learning_rate": 7.969739619985925e-05,
1405
+ "loss": 0.3664,
1406
+ "step": 3920
1407
+ },
1408
+ {
1409
+ "epoch": 1.274256144890039,
1410
+ "grad_norm": 0.3885388970375061,
1411
+ "learning_rate": 7.899366643209009e-05,
1412
+ "loss": 0.3415,
1413
+ "step": 3940
1414
+ },
1415
+ {
1416
+ "epoch": 1.2807244501940491,
1417
+ "grad_norm": 0.4131453335285187,
1418
+ "learning_rate": 7.82899366643209e-05,
1419
+ "loss": 0.3552,
1420
+ "step": 3960
1421
+ },
1422
+ {
1423
+ "epoch": 1.2871927554980596,
1424
+ "grad_norm": 0.39486396312713623,
1425
+ "learning_rate": 7.758620689655173e-05,
1426
+ "loss": 0.366,
1427
+ "step": 3980
1428
+ },
1429
+ {
1430
+ "epoch": 1.2936610608020698,
1431
+ "grad_norm": 0.3580794930458069,
1432
+ "learning_rate": 7.688247712878254e-05,
1433
+ "loss": 0.3585,
1434
+ "step": 4000
1435
+ },
1436
+ {
1437
+ "epoch": 1.2936610608020698,
1438
+ "eval_loss": 0.5952740907669067,
1439
+ "eval_runtime": 967.8131,
1440
+ "eval_samples_per_second": 6.331,
1441
+ "eval_steps_per_second": 0.791,
1442
+ "step": 4000
1443
+ }
1444
+ ],
1445
+ "logging_steps": 20,
1446
+ "max_steps": 6184,
1447
+ "num_input_tokens_seen": 0,
1448
+ "num_train_epochs": 2,
1449
+ "save_steps": 1000,
1450
+ "stateful_callbacks": {
1451
+ "TrainerControl": {
1452
+ "args": {
1453
+ "should_epoch_stop": false,
1454
+ "should_evaluate": false,
1455
+ "should_log": false,
1456
+ "should_save": true,
1457
+ "should_training_stop": false
1458
+ },
1459
+ "attributes": {}
1460
+ }
1461
+ },
1462
+ "total_flos": 5.33128270139818e+18,
1463
+ "train_batch_size": 8,
1464
+ "trial_name": null,
1465
+ "trial_params": null
1466
+ }
ultival_mistral/checkpoint-4000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8e5db938f70ff2ddfe787ade649d14bb55cc2be7a65bd63a20c3303bf68492
3
+ size 5368
ultival_mistral/checkpoint-5000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Ministral-8B-Instruct-2410
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
ultival_mistral/checkpoint-5000/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Ministral-8B-Instruct-2410",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 32,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "o_proj",
28
+ "k_proj",
29
+ "down_proj",
30
+ "q_proj",
31
+ "gate_proj",
32
+ "up_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
ultival_mistral/checkpoint-5000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d4e5d637b24ecf43d167a40f0933906906cec8c26ae36a745ac7a45e03349e3
3
+ size 349243752
ultival_mistral/checkpoint-5000/chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{% if messages[1]['role'] == 'user' %}{{ '[INST] ' + messages[0]['content'] + ' ' + messages[1]['content'] + ' [/INST]' }}{% set loop_messages = messages[2:] %}{% else %}{{ '[INST] ' + messages[0]['content'] + ' [/INST]' }}{% set loop_messages = messages[1:] %}{% endif %}{% else %}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}
ultival_mistral/checkpoint-5000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69e7979fd79b7efdd90afb4a0ff92abf491ca11ea63da3fa64e3f46b1d4b47b0
3
+ size 177919028
ultival_mistral/checkpoint-5000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:301bae54f3043b54f78ee582aa05b8bf01996d027ad704f9a95f10d8be516262
3
+ size 14244
ultival_mistral/checkpoint-5000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90de424fe588568c8644c48a645a04810ccb62dd370abc6f52d176fcc605dfa4
3
+ size 1064
ultival_mistral/checkpoint-5000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<pad>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
ultival_mistral/checkpoint-5000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be2965d3e435dbdef9376e5bdb3dcc2bfc7e271aa5152ae896cf53c7e920adf8
3
+ size 17078235
ultival_mistral/checkpoint-5000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
ultival_mistral/checkpoint-5000/trainer_state.json ADDED
@@ -0,0 +1,1824 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3000,
3
+ "best_metric": 0.5850913524627686,
4
+ "best_model_checkpoint": "./ultival_run2/checkpoint-3000",
5
+ "epoch": 1.6170763260025873,
6
+ "eval_steps": 1000,
7
+ "global_step": 5000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00646830530401035,
14
+ "grad_norm": 0.8976479172706604,
15
+ "learning_rate": 7.6e-06,
16
+ "loss": 1.5544,
17
+ "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.0129366106080207,
21
+ "grad_norm": 0.726784348487854,
22
+ "learning_rate": 1.56e-05,
23
+ "loss": 1.242,
24
+ "step": 40
25
+ },
26
+ {
27
+ "epoch": 0.019404915912031046,
28
+ "grad_norm": 0.4917074143886566,
29
+ "learning_rate": 2.36e-05,
30
+ "loss": 0.6694,
31
+ "step": 60
32
+ },
33
+ {
34
+ "epoch": 0.0258732212160414,
35
+ "grad_norm": 0.3471176326274872,
36
+ "learning_rate": 3.16e-05,
37
+ "loss": 0.5334,
38
+ "step": 80
39
+ },
40
+ {
41
+ "epoch": 0.03234152652005175,
42
+ "grad_norm": 0.41759616136550903,
43
+ "learning_rate": 3.960000000000001e-05,
44
+ "loss": 0.5435,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.03880983182406209,
49
+ "grad_norm": 0.3755040466785431,
50
+ "learning_rate": 4.76e-05,
51
+ "loss": 0.5684,
52
+ "step": 120
53
+ },
54
+ {
55
+ "epoch": 0.045278137128072445,
56
+ "grad_norm": 0.37058988213539124,
57
+ "learning_rate": 5.560000000000001e-05,
58
+ "loss": 0.5329,
59
+ "step": 140
60
+ },
61
+ {
62
+ "epoch": 0.0517464424320828,
63
+ "grad_norm": 0.42603132128715515,
64
+ "learning_rate": 6.36e-05,
65
+ "loss": 0.5457,
66
+ "step": 160
67
+ },
68
+ {
69
+ "epoch": 0.05821474773609314,
70
+ "grad_norm": 0.3154866099357605,
71
+ "learning_rate": 7.16e-05,
72
+ "loss": 0.5472,
73
+ "step": 180
74
+ },
75
+ {
76
+ "epoch": 0.0646830530401035,
77
+ "grad_norm": 0.3919835388660431,
78
+ "learning_rate": 7.960000000000001e-05,
79
+ "loss": 0.4675,
80
+ "step": 200
81
+ },
82
+ {
83
+ "epoch": 0.07115135834411385,
84
+ "grad_norm": 0.32180657982826233,
85
+ "learning_rate": 8.76e-05,
86
+ "loss": 0.5078,
87
+ "step": 220
88
+ },
89
+ {
90
+ "epoch": 0.07761966364812418,
91
+ "grad_norm": 0.3246982991695404,
92
+ "learning_rate": 9.56e-05,
93
+ "loss": 0.488,
94
+ "step": 240
95
+ },
96
+ {
97
+ "epoch": 0.08408796895213454,
98
+ "grad_norm": 0.341487318277359,
99
+ "learning_rate": 0.00010360000000000001,
100
+ "loss": 0.5058,
101
+ "step": 260
102
+ },
103
+ {
104
+ "epoch": 0.09055627425614489,
105
+ "grad_norm": 0.3597952127456665,
106
+ "learning_rate": 0.00011160000000000002,
107
+ "loss": 0.4574,
108
+ "step": 280
109
+ },
110
+ {
111
+ "epoch": 0.09702457956015524,
112
+ "grad_norm": 0.40912091732025146,
113
+ "learning_rate": 0.00011960000000000001,
114
+ "loss": 0.4931,
115
+ "step": 300
116
+ },
117
+ {
118
+ "epoch": 0.1034928848641656,
119
+ "grad_norm": 0.37430697679519653,
120
+ "learning_rate": 0.0001276,
121
+ "loss": 0.4724,
122
+ "step": 320
123
+ },
124
+ {
125
+ "epoch": 0.10996119016817593,
126
+ "grad_norm": 0.3920972943305969,
127
+ "learning_rate": 0.00013560000000000002,
128
+ "loss": 0.4997,
129
+ "step": 340
130
+ },
131
+ {
132
+ "epoch": 0.11642949547218628,
133
+ "grad_norm": 0.32511982321739197,
134
+ "learning_rate": 0.0001436,
135
+ "loss": 0.4889,
136
+ "step": 360
137
+ },
138
+ {
139
+ "epoch": 0.12289780077619664,
140
+ "grad_norm": 0.3926583528518677,
141
+ "learning_rate": 0.0001516,
142
+ "loss": 0.4772,
143
+ "step": 380
144
+ },
145
+ {
146
+ "epoch": 0.129366106080207,
147
+ "grad_norm": 0.3535909950733185,
148
+ "learning_rate": 0.0001596,
149
+ "loss": 0.465,
150
+ "step": 400
151
+ },
152
+ {
153
+ "epoch": 0.13583441138421734,
154
+ "grad_norm": 0.34381625056266785,
155
+ "learning_rate": 0.0001676,
156
+ "loss": 0.4934,
157
+ "step": 420
158
+ },
159
+ {
160
+ "epoch": 0.1423027166882277,
161
+ "grad_norm": 0.3753703236579895,
162
+ "learning_rate": 0.0001756,
163
+ "loss": 0.5068,
164
+ "step": 440
165
+ },
166
+ {
167
+ "epoch": 0.14877102199223805,
168
+ "grad_norm": 0.34895479679107666,
169
+ "learning_rate": 0.00018360000000000002,
170
+ "loss": 0.5186,
171
+ "step": 460
172
+ },
173
+ {
174
+ "epoch": 0.15523932729624837,
175
+ "grad_norm": 0.24720332026481628,
176
+ "learning_rate": 0.0001916,
177
+ "loss": 0.461,
178
+ "step": 480
179
+ },
180
+ {
181
+ "epoch": 0.16170763260025872,
182
+ "grad_norm": 0.3713696002960205,
183
+ "learning_rate": 0.0001996,
184
+ "loss": 0.4932,
185
+ "step": 500
186
+ },
187
+ {
188
+ "epoch": 0.16817593790426907,
189
+ "grad_norm": 0.3724994659423828,
190
+ "learning_rate": 0.00019933145672061927,
191
+ "loss": 0.5036,
192
+ "step": 520
193
+ },
194
+ {
195
+ "epoch": 0.17464424320827943,
196
+ "grad_norm": 0.3562956750392914,
197
+ "learning_rate": 0.00019862772695285012,
198
+ "loss": 0.4871,
199
+ "step": 540
200
+ },
201
+ {
202
+ "epoch": 0.18111254851228978,
203
+ "grad_norm": 0.331737756729126,
204
+ "learning_rate": 0.00019792399718508095,
205
+ "loss": 0.5178,
206
+ "step": 560
207
+ },
208
+ {
209
+ "epoch": 0.18758085381630013,
210
+ "grad_norm": 0.34049859642982483,
211
+ "learning_rate": 0.00019722026741731178,
212
+ "loss": 0.488,
213
+ "step": 580
214
+ },
215
+ {
216
+ "epoch": 0.19404915912031048,
217
+ "grad_norm": 0.36076661944389343,
218
+ "learning_rate": 0.00019651653764954258,
219
+ "loss": 0.4821,
220
+ "step": 600
221
+ },
222
+ {
223
+ "epoch": 0.20051746442432083,
224
+ "grad_norm": 0.3059399127960205,
225
+ "learning_rate": 0.0001958128078817734,
226
+ "loss": 0.4924,
227
+ "step": 620
228
+ },
229
+ {
230
+ "epoch": 0.2069857697283312,
231
+ "grad_norm": 0.30692920088768005,
232
+ "learning_rate": 0.00019510907811400423,
233
+ "loss": 0.4822,
234
+ "step": 640
235
+ },
236
+ {
237
+ "epoch": 0.21345407503234154,
238
+ "grad_norm": 0.3319448232650757,
239
+ "learning_rate": 0.00019440534834623506,
240
+ "loss": 0.4607,
241
+ "step": 660
242
+ },
243
+ {
244
+ "epoch": 0.21992238033635186,
245
+ "grad_norm": 0.3047769069671631,
246
+ "learning_rate": 0.00019370161857846588,
247
+ "loss": 0.5357,
248
+ "step": 680
249
+ },
250
+ {
251
+ "epoch": 0.22639068564036222,
252
+ "grad_norm": 0.4353258013725281,
253
+ "learning_rate": 0.00019299788881069668,
254
+ "loss": 0.4738,
255
+ "step": 700
256
+ },
257
+ {
258
+ "epoch": 0.23285899094437257,
259
+ "grad_norm": 0.30877336859703064,
260
+ "learning_rate": 0.00019229415904292754,
261
+ "loss": 0.5128,
262
+ "step": 720
263
+ },
264
+ {
265
+ "epoch": 0.23932729624838292,
266
+ "grad_norm": 0.3188530206680298,
267
+ "learning_rate": 0.00019159042927515836,
268
+ "loss": 0.4651,
269
+ "step": 740
270
+ },
271
+ {
272
+ "epoch": 0.24579560155239327,
273
+ "grad_norm": 0.40678778290748596,
274
+ "learning_rate": 0.0001908866995073892,
275
+ "loss": 0.5112,
276
+ "step": 760
277
+ },
278
+ {
279
+ "epoch": 0.2522639068564036,
280
+ "grad_norm": 0.34188327193260193,
281
+ "learning_rate": 0.00019018296973962,
282
+ "loss": 0.487,
283
+ "step": 780
284
+ },
285
+ {
286
+ "epoch": 0.258732212160414,
287
+ "grad_norm": 0.3002854585647583,
288
+ "learning_rate": 0.00018947923997185081,
289
+ "loss": 0.512,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.2652005174644243,
294
+ "grad_norm": 0.35756272077560425,
295
+ "learning_rate": 0.00018877551020408164,
296
+ "loss": 0.4812,
297
+ "step": 820
298
+ },
299
+ {
300
+ "epoch": 0.2716688227684347,
301
+ "grad_norm": 0.33527258038520813,
302
+ "learning_rate": 0.00018807178043631247,
303
+ "loss": 0.4895,
304
+ "step": 840
305
+ },
306
+ {
307
+ "epoch": 0.278137128072445,
308
+ "grad_norm": 0.3472883105278015,
309
+ "learning_rate": 0.0001873680506685433,
310
+ "loss": 0.4671,
311
+ "step": 860
312
+ },
313
+ {
314
+ "epoch": 0.2846054333764554,
315
+ "grad_norm": 0.3604671061038971,
316
+ "learning_rate": 0.0001866643209007741,
317
+ "loss": 0.5233,
318
+ "step": 880
319
+ },
320
+ {
321
+ "epoch": 0.2910737386804657,
322
+ "grad_norm": 0.31021153926849365,
323
+ "learning_rate": 0.00018596059113300492,
324
+ "loss": 0.491,
325
+ "step": 900
326
+ },
327
+ {
328
+ "epoch": 0.2975420439844761,
329
+ "grad_norm": 0.6417067050933838,
330
+ "learning_rate": 0.00018525686136523577,
331
+ "loss": 0.4858,
332
+ "step": 920
333
+ },
334
+ {
335
+ "epoch": 0.3040103492884864,
336
+ "grad_norm": 0.33719781041145325,
337
+ "learning_rate": 0.0001845531315974666,
338
+ "loss": 0.4722,
339
+ "step": 940
340
+ },
341
+ {
342
+ "epoch": 0.31047865459249674,
343
+ "grad_norm": 0.3568169176578522,
344
+ "learning_rate": 0.0001838494018296974,
345
+ "loss": 0.5088,
346
+ "step": 960
347
+ },
348
+ {
349
+ "epoch": 0.3169469598965071,
350
+ "grad_norm": 0.42162615060806274,
351
+ "learning_rate": 0.00018314567206192823,
352
+ "loss": 0.5091,
353
+ "step": 980
354
+ },
355
+ {
356
+ "epoch": 0.32341526520051744,
357
+ "grad_norm": 0.3212260603904724,
358
+ "learning_rate": 0.00018244194229415905,
359
+ "loss": 0.4654,
360
+ "step": 1000
361
+ },
362
+ {
363
+ "epoch": 0.32341526520051744,
364
+ "eval_loss": 0.6167479753494263,
365
+ "eval_runtime": 1013.3659,
366
+ "eval_samples_per_second": 6.046,
367
+ "eval_steps_per_second": 0.756,
368
+ "step": 1000
369
+ },
370
+ {
371
+ "epoch": 0.3298835705045278,
372
+ "grad_norm": 0.3184376060962677,
373
+ "learning_rate": 0.00018173821252638988,
374
+ "loss": 0.4712,
375
+ "step": 1020
376
+ },
377
+ {
378
+ "epoch": 0.33635187580853815,
379
+ "grad_norm": 0.3114645183086395,
380
+ "learning_rate": 0.0001810344827586207,
381
+ "loss": 0.4393,
382
+ "step": 1040
383
+ },
384
+ {
385
+ "epoch": 0.3428201811125485,
386
+ "grad_norm": 0.30122488737106323,
387
+ "learning_rate": 0.0001803307529908515,
388
+ "loss": 0.4882,
389
+ "step": 1060
390
+ },
391
+ {
392
+ "epoch": 0.34928848641655885,
393
+ "grad_norm": 0.2877393066883087,
394
+ "learning_rate": 0.00017962702322308233,
395
+ "loss": 0.413,
396
+ "step": 1080
397
+ },
398
+ {
399
+ "epoch": 0.35575679172056923,
400
+ "grad_norm": 0.33227887749671936,
401
+ "learning_rate": 0.00017892329345531319,
402
+ "loss": 0.4527,
403
+ "step": 1100
404
+ },
405
+ {
406
+ "epoch": 0.36222509702457956,
407
+ "grad_norm": 0.33423128724098206,
408
+ "learning_rate": 0.000178219563687544,
409
+ "loss": 0.4604,
410
+ "step": 1120
411
+ },
412
+ {
413
+ "epoch": 0.36869340232858994,
414
+ "grad_norm": 0.36922487616539,
415
+ "learning_rate": 0.0001775158339197748,
416
+ "loss": 0.5044,
417
+ "step": 1140
418
+ },
419
+ {
420
+ "epoch": 0.37516170763260026,
421
+ "grad_norm": 0.34816262125968933,
422
+ "learning_rate": 0.00017681210415200564,
423
+ "loss": 0.4898,
424
+ "step": 1160
425
+ },
426
+ {
427
+ "epoch": 0.3816300129366106,
428
+ "grad_norm": 0.32086044549942017,
429
+ "learning_rate": 0.00017610837438423646,
430
+ "loss": 0.4647,
431
+ "step": 1180
432
+ },
433
+ {
434
+ "epoch": 0.38809831824062097,
435
+ "grad_norm": 0.3879236876964569,
436
+ "learning_rate": 0.0001754046446164673,
437
+ "loss": 0.5179,
438
+ "step": 1200
439
+ },
440
+ {
441
+ "epoch": 0.3945666235446313,
442
+ "grad_norm": 0.2899722158908844,
443
+ "learning_rate": 0.00017470091484869812,
444
+ "loss": 0.4786,
445
+ "step": 1220
446
+ },
447
+ {
448
+ "epoch": 0.40103492884864167,
449
+ "grad_norm": 0.3844458758831024,
450
+ "learning_rate": 0.00017399718508092892,
451
+ "loss": 0.4513,
452
+ "step": 1240
453
+ },
454
+ {
455
+ "epoch": 0.407503234152652,
456
+ "grad_norm": 0.3678169548511505,
457
+ "learning_rate": 0.00017329345531315974,
458
+ "loss": 0.4789,
459
+ "step": 1260
460
+ },
461
+ {
462
+ "epoch": 0.4139715394566624,
463
+ "grad_norm": 0.37138116359710693,
464
+ "learning_rate": 0.00017258972554539057,
465
+ "loss": 0.4694,
466
+ "step": 1280
467
+ },
468
+ {
469
+ "epoch": 0.4204398447606727,
470
+ "grad_norm": 0.3604221045970917,
471
+ "learning_rate": 0.00017188599577762142,
472
+ "loss": 0.4496,
473
+ "step": 1300
474
+ },
475
+ {
476
+ "epoch": 0.4269081500646831,
477
+ "grad_norm": 0.33961811661720276,
478
+ "learning_rate": 0.00017118226600985222,
479
+ "loss": 0.4696,
480
+ "step": 1320
481
+ },
482
+ {
483
+ "epoch": 0.4333764553686934,
484
+ "grad_norm": 0.3244430124759674,
485
+ "learning_rate": 0.00017047853624208305,
486
+ "loss": 0.4322,
487
+ "step": 1340
488
+ },
489
+ {
490
+ "epoch": 0.4398447606727037,
491
+ "grad_norm": 0.2879267930984497,
492
+ "learning_rate": 0.00016977480647431388,
493
+ "loss": 0.4646,
494
+ "step": 1360
495
+ },
496
+ {
497
+ "epoch": 0.4463130659767141,
498
+ "grad_norm": 0.3437345027923584,
499
+ "learning_rate": 0.0001690710767065447,
500
+ "loss": 0.4711,
501
+ "step": 1380
502
+ },
503
+ {
504
+ "epoch": 0.45278137128072443,
505
+ "grad_norm": 0.3166882395744324,
506
+ "learning_rate": 0.00016836734693877553,
507
+ "loss": 0.486,
508
+ "step": 1400
509
+ },
510
+ {
511
+ "epoch": 0.4592496765847348,
512
+ "grad_norm": 0.3677555322647095,
513
+ "learning_rate": 0.00016766361717100633,
514
+ "loss": 0.4677,
515
+ "step": 1420
516
+ },
517
+ {
518
+ "epoch": 0.46571798188874514,
519
+ "grad_norm": 0.32657769322395325,
520
+ "learning_rate": 0.00016695988740323715,
521
+ "loss": 0.4649,
522
+ "step": 1440
523
+ },
524
+ {
525
+ "epoch": 0.4721862871927555,
526
+ "grad_norm": 0.2980731427669525,
527
+ "learning_rate": 0.00016625615763546798,
528
+ "loss": 0.4547,
529
+ "step": 1460
530
+ },
531
+ {
532
+ "epoch": 0.47865459249676584,
533
+ "grad_norm": 0.3690549433231354,
534
+ "learning_rate": 0.0001655524278676988,
535
+ "loss": 0.4512,
536
+ "step": 1480
537
+ },
538
+ {
539
+ "epoch": 0.4851228978007762,
540
+ "grad_norm": 0.31456947326660156,
541
+ "learning_rate": 0.00016484869809992963,
542
+ "loss": 0.4675,
543
+ "step": 1500
544
+ },
545
+ {
546
+ "epoch": 0.49159120310478654,
547
+ "grad_norm": 0.3823802173137665,
548
+ "learning_rate": 0.00016414496833216046,
549
+ "loss": 0.4849,
550
+ "step": 1520
551
+ },
552
+ {
553
+ "epoch": 0.49805950840879687,
554
+ "grad_norm": 0.33280324935913086,
555
+ "learning_rate": 0.0001634412385643913,
556
+ "loss": 0.4876,
557
+ "step": 1540
558
+ },
559
+ {
560
+ "epoch": 0.5045278137128072,
561
+ "grad_norm": 0.28788062930107117,
562
+ "learning_rate": 0.00016273750879662211,
563
+ "loss": 0.426,
564
+ "step": 1560
565
+ },
566
+ {
567
+ "epoch": 0.5109961190168176,
568
+ "grad_norm": 0.3809375762939453,
569
+ "learning_rate": 0.00016203377902885294,
570
+ "loss": 0.4676,
571
+ "step": 1580
572
+ },
573
+ {
574
+ "epoch": 0.517464424320828,
575
+ "grad_norm": 0.27940312027931213,
576
+ "learning_rate": 0.00016133004926108374,
577
+ "loss": 0.4413,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 0.5239327296248383,
582
+ "grad_norm": 0.30946919322013855,
583
+ "learning_rate": 0.00016062631949331457,
584
+ "loss": 0.4395,
585
+ "step": 1620
586
+ },
587
+ {
588
+ "epoch": 0.5304010349288486,
589
+ "grad_norm": 0.36931583285331726,
590
+ "learning_rate": 0.0001599225897255454,
591
+ "loss": 0.4565,
592
+ "step": 1640
593
+ },
594
+ {
595
+ "epoch": 0.536869340232859,
596
+ "grad_norm": 0.3652470111846924,
597
+ "learning_rate": 0.00015921885995777622,
598
+ "loss": 0.4522,
599
+ "step": 1660
600
+ },
601
+ {
602
+ "epoch": 0.5433376455368694,
603
+ "grad_norm": 0.3060782551765442,
604
+ "learning_rate": 0.00015851513019000705,
605
+ "loss": 0.45,
606
+ "step": 1680
607
+ },
608
+ {
609
+ "epoch": 0.5498059508408797,
610
+ "grad_norm": 0.3275803029537201,
611
+ "learning_rate": 0.00015781140042223787,
612
+ "loss": 0.445,
613
+ "step": 1700
614
+ },
615
+ {
616
+ "epoch": 0.55627425614489,
617
+ "grad_norm": 0.34865352511405945,
618
+ "learning_rate": 0.0001571076706544687,
619
+ "loss": 0.4306,
620
+ "step": 1720
621
+ },
622
+ {
623
+ "epoch": 0.5627425614489003,
624
+ "grad_norm": 0.42525383830070496,
625
+ "learning_rate": 0.00015640394088669953,
626
+ "loss": 0.4544,
627
+ "step": 1740
628
+ },
629
+ {
630
+ "epoch": 0.5692108667529108,
631
+ "grad_norm": 0.32165050506591797,
632
+ "learning_rate": 0.00015570021111893035,
633
+ "loss": 0.484,
634
+ "step": 1760
635
+ },
636
+ {
637
+ "epoch": 0.5756791720569211,
638
+ "grad_norm": 0.3213440477848053,
639
+ "learning_rate": 0.00015499648135116115,
640
+ "loss": 0.4469,
641
+ "step": 1780
642
+ },
643
+ {
644
+ "epoch": 0.5821474773609314,
645
+ "grad_norm": 0.42835432291030884,
646
+ "learning_rate": 0.00015429275158339198,
647
+ "loss": 0.4592,
648
+ "step": 1800
649
+ },
650
+ {
651
+ "epoch": 0.5886157826649417,
652
+ "grad_norm": 0.3861497938632965,
653
+ "learning_rate": 0.0001535890218156228,
654
+ "loss": 0.4607,
655
+ "step": 1820
656
+ },
657
+ {
658
+ "epoch": 0.5950840879689522,
659
+ "grad_norm": 0.4156148135662079,
660
+ "learning_rate": 0.00015288529204785363,
661
+ "loss": 0.4541,
662
+ "step": 1840
663
+ },
664
+ {
665
+ "epoch": 0.6015523932729625,
666
+ "grad_norm": 0.32371824979782104,
667
+ "learning_rate": 0.00015218156228008446,
668
+ "loss": 0.4287,
669
+ "step": 1860
670
+ },
671
+ {
672
+ "epoch": 0.6080206985769728,
673
+ "grad_norm": 0.363788366317749,
674
+ "learning_rate": 0.00015147783251231528,
675
+ "loss": 0.428,
676
+ "step": 1880
677
+ },
678
+ {
679
+ "epoch": 0.6144890038809832,
680
+ "grad_norm": 0.3932839334011078,
681
+ "learning_rate": 0.0001507741027445461,
682
+ "loss": 0.4359,
683
+ "step": 1900
684
+ },
685
+ {
686
+ "epoch": 0.6209573091849935,
687
+ "grad_norm": 0.3009307086467743,
688
+ "learning_rate": 0.00015007037297677694,
689
+ "loss": 0.3976,
690
+ "step": 1920
691
+ },
692
+ {
693
+ "epoch": 0.6274256144890039,
694
+ "grad_norm": 0.3484659194946289,
695
+ "learning_rate": 0.00014936664320900776,
696
+ "loss": 0.4584,
697
+ "step": 1940
698
+ },
699
+ {
700
+ "epoch": 0.6338939197930142,
701
+ "grad_norm": 0.33728504180908203,
702
+ "learning_rate": 0.00014866291344123856,
703
+ "loss": 0.4543,
704
+ "step": 1960
705
+ },
706
+ {
707
+ "epoch": 0.6403622250970246,
708
+ "grad_norm": 0.3207024037837982,
709
+ "learning_rate": 0.0001479591836734694,
710
+ "loss": 0.4611,
711
+ "step": 1980
712
+ },
713
+ {
714
+ "epoch": 0.6468305304010349,
715
+ "grad_norm": 0.29134076833724976,
716
+ "learning_rate": 0.00014725545390570022,
717
+ "loss": 0.439,
718
+ "step": 2000
719
+ },
720
+ {
721
+ "epoch": 0.6468305304010349,
722
+ "eval_loss": 0.6003134250640869,
723
+ "eval_runtime": 968.0302,
724
+ "eval_samples_per_second": 6.329,
725
+ "eval_steps_per_second": 0.791,
726
+ "step": 2000
727
+ },
728
+ {
729
+ "epoch": 0.6532988357050453,
730
+ "grad_norm": 0.3548874258995056,
731
+ "learning_rate": 0.00014655172413793104,
732
+ "loss": 0.4548,
733
+ "step": 2020
734
+ },
735
+ {
736
+ "epoch": 0.6597671410090556,
737
+ "grad_norm": 0.3707274794578552,
738
+ "learning_rate": 0.00014584799437016187,
739
+ "loss": 0.4271,
740
+ "step": 2040
741
+ },
742
+ {
743
+ "epoch": 0.666235446313066,
744
+ "grad_norm": 0.37172695994377136,
745
+ "learning_rate": 0.0001451442646023927,
746
+ "loss": 0.4626,
747
+ "step": 2060
748
+ },
749
+ {
750
+ "epoch": 0.6727037516170763,
751
+ "grad_norm": 0.3459303081035614,
752
+ "learning_rate": 0.00014444053483462352,
753
+ "loss": 0.4317,
754
+ "step": 2080
755
+ },
756
+ {
757
+ "epoch": 0.6791720569210866,
758
+ "grad_norm": 0.3477213680744171,
759
+ "learning_rate": 0.00014373680506685435,
760
+ "loss": 0.4243,
761
+ "step": 2100
762
+ },
763
+ {
764
+ "epoch": 0.685640362225097,
765
+ "grad_norm": 0.31847599148750305,
766
+ "learning_rate": 0.00014303307529908517,
767
+ "loss": 0.4234,
768
+ "step": 2120
769
+ },
770
+ {
771
+ "epoch": 0.6921086675291074,
772
+ "grad_norm": 0.3231470584869385,
773
+ "learning_rate": 0.00014232934553131597,
774
+ "loss": 0.4302,
775
+ "step": 2140
776
+ },
777
+ {
778
+ "epoch": 0.6985769728331177,
779
+ "grad_norm": 0.3745759427547455,
780
+ "learning_rate": 0.0001416256157635468,
781
+ "loss": 0.4533,
782
+ "step": 2160
783
+ },
784
+ {
785
+ "epoch": 0.705045278137128,
786
+ "grad_norm": 0.34834805130958557,
787
+ "learning_rate": 0.00014092188599577763,
788
+ "loss": 0.4748,
789
+ "step": 2180
790
+ },
791
+ {
792
+ "epoch": 0.7115135834411385,
793
+ "grad_norm": 0.3675827085971832,
794
+ "learning_rate": 0.00014021815622800845,
795
+ "loss": 0.4401,
796
+ "step": 2200
797
+ },
798
+ {
799
+ "epoch": 0.7179818887451488,
800
+ "grad_norm": 0.2839801013469696,
801
+ "learning_rate": 0.00013951442646023928,
802
+ "loss": 0.4332,
803
+ "step": 2220
804
+ },
805
+ {
806
+ "epoch": 0.7244501940491591,
807
+ "grad_norm": 0.27037495374679565,
808
+ "learning_rate": 0.00013881069669247008,
809
+ "loss": 0.4382,
810
+ "step": 2240
811
+ },
812
+ {
813
+ "epoch": 0.7309184993531694,
814
+ "grad_norm": 0.34917208552360535,
815
+ "learning_rate": 0.00013810696692470093,
816
+ "loss": 0.4251,
817
+ "step": 2260
818
+ },
819
+ {
820
+ "epoch": 0.7373868046571799,
821
+ "grad_norm": 0.39801281690597534,
822
+ "learning_rate": 0.00013740323715693176,
823
+ "loss": 0.4432,
824
+ "step": 2280
825
+ },
826
+ {
827
+ "epoch": 0.7438551099611902,
828
+ "grad_norm": 0.8932256102561951,
829
+ "learning_rate": 0.00013669950738916259,
830
+ "loss": 0.4577,
831
+ "step": 2300
832
+ },
833
+ {
834
+ "epoch": 0.7503234152652005,
835
+ "grad_norm": 0.35125109553337097,
836
+ "learning_rate": 0.00013599577762139339,
837
+ "loss": 0.4502,
838
+ "step": 2320
839
+ },
840
+ {
841
+ "epoch": 0.7567917205692108,
842
+ "grad_norm": 0.3012046813964844,
843
+ "learning_rate": 0.0001352920478536242,
844
+ "loss": 0.4513,
845
+ "step": 2340
846
+ },
847
+ {
848
+ "epoch": 0.7632600258732212,
849
+ "grad_norm": 0.3242063820362091,
850
+ "learning_rate": 0.00013458831808585504,
851
+ "loss": 0.4559,
852
+ "step": 2360
853
+ },
854
+ {
855
+ "epoch": 0.7697283311772316,
856
+ "grad_norm": 0.3523860275745392,
857
+ "learning_rate": 0.00013388458831808587,
858
+ "loss": 0.4434,
859
+ "step": 2380
860
+ },
861
+ {
862
+ "epoch": 0.7761966364812419,
863
+ "grad_norm": 0.29160013794898987,
864
+ "learning_rate": 0.0001331808585503167,
865
+ "loss": 0.44,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 0.7826649417852523,
870
+ "grad_norm": 0.3047971725463867,
871
+ "learning_rate": 0.0001324771287825475,
872
+ "loss": 0.4336,
873
+ "step": 2420
874
+ },
875
+ {
876
+ "epoch": 0.7891332470892626,
877
+ "grad_norm": 0.3192126750946045,
878
+ "learning_rate": 0.00013177339901477832,
879
+ "loss": 0.4495,
880
+ "step": 2440
881
+ },
882
+ {
883
+ "epoch": 0.795601552393273,
884
+ "grad_norm": 0.37997299432754517,
885
+ "learning_rate": 0.00013106966924700917,
886
+ "loss": 0.4685,
887
+ "step": 2460
888
+ },
889
+ {
890
+ "epoch": 0.8020698576972833,
891
+ "grad_norm": 0.2811839282512665,
892
+ "learning_rate": 0.00013036593947924,
893
+ "loss": 0.4206,
894
+ "step": 2480
895
+ },
896
+ {
897
+ "epoch": 0.8085381630012937,
898
+ "grad_norm": 0.3121369779109955,
899
+ "learning_rate": 0.0001296622097114708,
900
+ "loss": 0.425,
901
+ "step": 2500
902
+ },
903
+ {
904
+ "epoch": 0.815006468305304,
905
+ "grad_norm": 0.3134746253490448,
906
+ "learning_rate": 0.00012895847994370162,
907
+ "loss": 0.4498,
908
+ "step": 2520
909
+ },
910
+ {
911
+ "epoch": 0.8214747736093143,
912
+ "grad_norm": 0.36752375960350037,
913
+ "learning_rate": 0.00012825475017593245,
914
+ "loss": 0.4288,
915
+ "step": 2540
916
+ },
917
+ {
918
+ "epoch": 0.8279430789133247,
919
+ "grad_norm": 0.3915141224861145,
920
+ "learning_rate": 0.00012755102040816328,
921
+ "loss": 0.4605,
922
+ "step": 2560
923
+ },
924
+ {
925
+ "epoch": 0.8344113842173351,
926
+ "grad_norm": 0.3662431538105011,
927
+ "learning_rate": 0.00012684729064039408,
928
+ "loss": 0.4194,
929
+ "step": 2580
930
+ },
931
+ {
932
+ "epoch": 0.8408796895213454,
933
+ "grad_norm": 0.3387574851512909,
934
+ "learning_rate": 0.0001261435608726249,
935
+ "loss": 0.4258,
936
+ "step": 2600
937
+ },
938
+ {
939
+ "epoch": 0.8473479948253557,
940
+ "grad_norm": 0.38025060296058655,
941
+ "learning_rate": 0.00012543983110485573,
942
+ "loss": 0.4291,
943
+ "step": 2620
944
+ },
945
+ {
946
+ "epoch": 0.8538163001293662,
947
+ "grad_norm": 0.3517441749572754,
948
+ "learning_rate": 0.00012473610133708658,
949
+ "loss": 0.4384,
950
+ "step": 2640
951
+ },
952
+ {
953
+ "epoch": 0.8602846054333765,
954
+ "grad_norm": 0.33037152886390686,
955
+ "learning_rate": 0.0001240323715693174,
956
+ "loss": 0.4452,
957
+ "step": 2660
958
+ },
959
+ {
960
+ "epoch": 0.8667529107373868,
961
+ "grad_norm": 0.3645654618740082,
962
+ "learning_rate": 0.0001233286418015482,
963
+ "loss": 0.3992,
964
+ "step": 2680
965
+ },
966
+ {
967
+ "epoch": 0.8732212160413971,
968
+ "grad_norm": 0.32396814227104187,
969
+ "learning_rate": 0.00012262491203377904,
970
+ "loss": 0.4,
971
+ "step": 2700
972
+ },
973
+ {
974
+ "epoch": 0.8796895213454075,
975
+ "grad_norm": 0.3377726078033447,
976
+ "learning_rate": 0.00012192118226600986,
977
+ "loss": 0.4306,
978
+ "step": 2720
979
+ },
980
+ {
981
+ "epoch": 0.8861578266494179,
982
+ "grad_norm": 0.32345983386039734,
983
+ "learning_rate": 0.00012121745249824067,
984
+ "loss": 0.4225,
985
+ "step": 2740
986
+ },
987
+ {
988
+ "epoch": 0.8926261319534282,
989
+ "grad_norm": 0.34778884053230286,
990
+ "learning_rate": 0.0001205137227304715,
991
+ "loss": 0.4256,
992
+ "step": 2760
993
+ },
994
+ {
995
+ "epoch": 0.8990944372574385,
996
+ "grad_norm": 0.3861066997051239,
997
+ "learning_rate": 0.00011980999296270233,
998
+ "loss": 0.4465,
999
+ "step": 2780
1000
+ },
1001
+ {
1002
+ "epoch": 0.9055627425614489,
1003
+ "grad_norm": 0.3882784843444824,
1004
+ "learning_rate": 0.00011910626319493314,
1005
+ "loss": 0.4026,
1006
+ "step": 2800
1007
+ },
1008
+ {
1009
+ "epoch": 0.9120310478654593,
1010
+ "grad_norm": 0.4380407929420471,
1011
+ "learning_rate": 0.00011840253342716397,
1012
+ "loss": 0.4375,
1013
+ "step": 2820
1014
+ },
1015
+ {
1016
+ "epoch": 0.9184993531694696,
1017
+ "grad_norm": 0.35943394899368286,
1018
+ "learning_rate": 0.00011769880365939481,
1019
+ "loss": 0.4337,
1020
+ "step": 2840
1021
+ },
1022
+ {
1023
+ "epoch": 0.92496765847348,
1024
+ "grad_norm": 0.3862978219985962,
1025
+ "learning_rate": 0.00011699507389162563,
1026
+ "loss": 0.3949,
1027
+ "step": 2860
1028
+ },
1029
+ {
1030
+ "epoch": 0.9314359637774903,
1031
+ "grad_norm": 0.34592369198799133,
1032
+ "learning_rate": 0.00011629134412385645,
1033
+ "loss": 0.3798,
1034
+ "step": 2880
1035
+ },
1036
+ {
1037
+ "epoch": 0.9379042690815006,
1038
+ "grad_norm": 0.34942877292633057,
1039
+ "learning_rate": 0.00011558761435608727,
1040
+ "loss": 0.4235,
1041
+ "step": 2900
1042
+ },
1043
+ {
1044
+ "epoch": 0.944372574385511,
1045
+ "grad_norm": 0.28018903732299805,
1046
+ "learning_rate": 0.00011488388458831809,
1047
+ "loss": 0.4353,
1048
+ "step": 2920
1049
+ },
1050
+ {
1051
+ "epoch": 0.9508408796895214,
1052
+ "grad_norm": 0.3505672812461853,
1053
+ "learning_rate": 0.00011418015482054891,
1054
+ "loss": 0.3954,
1055
+ "step": 2940
1056
+ },
1057
+ {
1058
+ "epoch": 0.9573091849935317,
1059
+ "grad_norm": 0.3264220654964447,
1060
+ "learning_rate": 0.00011347642505277974,
1061
+ "loss": 0.4422,
1062
+ "step": 2960
1063
+ },
1064
+ {
1065
+ "epoch": 0.963777490297542,
1066
+ "grad_norm": 0.36382535099983215,
1067
+ "learning_rate": 0.00011277269528501055,
1068
+ "loss": 0.4116,
1069
+ "step": 2980
1070
+ },
1071
+ {
1072
+ "epoch": 0.9702457956015524,
1073
+ "grad_norm": 0.3051494359970093,
1074
+ "learning_rate": 0.00011206896551724138,
1075
+ "loss": 0.4135,
1076
+ "step": 3000
1077
+ },
1078
+ {
1079
+ "epoch": 0.9702457956015524,
1080
+ "eval_loss": 0.5850913524627686,
1081
+ "eval_runtime": 967.6256,
1082
+ "eval_samples_per_second": 6.332,
1083
+ "eval_steps_per_second": 0.792,
1084
+ "step": 3000
1085
+ },
1086
+ {
1087
+ "epoch": 0.9767141009055628,
1088
+ "grad_norm": 0.34533271193504333,
1089
+ "learning_rate": 0.00011136523574947222,
1090
+ "loss": 0.4312,
1091
+ "step": 3020
1092
+ },
1093
+ {
1094
+ "epoch": 0.9831824062095731,
1095
+ "grad_norm": 0.35221609473228455,
1096
+ "learning_rate": 0.00011066150598170305,
1097
+ "loss": 0.4245,
1098
+ "step": 3040
1099
+ },
1100
+ {
1101
+ "epoch": 0.9896507115135834,
1102
+ "grad_norm": 0.3893713057041168,
1103
+ "learning_rate": 0.00010995777621393386,
1104
+ "loss": 0.4048,
1105
+ "step": 3060
1106
+ },
1107
+ {
1108
+ "epoch": 0.9961190168175937,
1109
+ "grad_norm": 0.3586278557777405,
1110
+ "learning_rate": 0.00010925404644616468,
1111
+ "loss": 0.4159,
1112
+ "step": 3080
1113
+ },
1114
+ {
1115
+ "epoch": 1.0025873221216042,
1116
+ "grad_norm": 0.32620444893836975,
1117
+ "learning_rate": 0.0001085503166783955,
1118
+ "loss": 0.3741,
1119
+ "step": 3100
1120
+ },
1121
+ {
1122
+ "epoch": 1.0090556274256144,
1123
+ "grad_norm": 0.33812645077705383,
1124
+ "learning_rate": 0.00010784658691062632,
1125
+ "loss": 0.3694,
1126
+ "step": 3120
1127
+ },
1128
+ {
1129
+ "epoch": 1.0155239327296248,
1130
+ "grad_norm": 0.39038708806037903,
1131
+ "learning_rate": 0.00010714285714285715,
1132
+ "loss": 0.3882,
1133
+ "step": 3140
1134
+ },
1135
+ {
1136
+ "epoch": 1.0219922380336353,
1137
+ "grad_norm": 0.3209969997406006,
1138
+ "learning_rate": 0.00010643912737508796,
1139
+ "loss": 0.3437,
1140
+ "step": 3160
1141
+ },
1142
+ {
1143
+ "epoch": 1.0284605433376455,
1144
+ "grad_norm": 0.3705078661441803,
1145
+ "learning_rate": 0.00010573539760731879,
1146
+ "loss": 0.3568,
1147
+ "step": 3180
1148
+ },
1149
+ {
1150
+ "epoch": 1.034928848641656,
1151
+ "grad_norm": 0.3562203645706177,
1152
+ "learning_rate": 0.0001050316678395496,
1153
+ "loss": 0.3916,
1154
+ "step": 3200
1155
+ },
1156
+ {
1157
+ "epoch": 1.0413971539456663,
1158
+ "grad_norm": 0.3724074363708496,
1159
+ "learning_rate": 0.00010432793807178046,
1160
+ "loss": 0.3866,
1161
+ "step": 3220
1162
+ },
1163
+ {
1164
+ "epoch": 1.0478654592496766,
1165
+ "grad_norm": 0.40453559160232544,
1166
+ "learning_rate": 0.00010362420830401127,
1167
+ "loss": 0.368,
1168
+ "step": 3240
1169
+ },
1170
+ {
1171
+ "epoch": 1.054333764553687,
1172
+ "grad_norm": 0.3316144049167633,
1173
+ "learning_rate": 0.0001029204785362421,
1174
+ "loss": 0.3696,
1175
+ "step": 3260
1176
+ },
1177
+ {
1178
+ "epoch": 1.0608020698576972,
1179
+ "grad_norm": 0.3503650724887848,
1180
+ "learning_rate": 0.00010221674876847291,
1181
+ "loss": 0.3644,
1182
+ "step": 3280
1183
+ },
1184
+ {
1185
+ "epoch": 1.0672703751617076,
1186
+ "grad_norm": 0.3136923909187317,
1187
+ "learning_rate": 0.00010151301900070374,
1188
+ "loss": 0.3414,
1189
+ "step": 3300
1190
+ },
1191
+ {
1192
+ "epoch": 1.073738680465718,
1193
+ "grad_norm": 0.3573864996433258,
1194
+ "learning_rate": 0.00010080928923293456,
1195
+ "loss": 0.3586,
1196
+ "step": 3320
1197
+ },
1198
+ {
1199
+ "epoch": 1.0802069857697283,
1200
+ "grad_norm": 0.26215964555740356,
1201
+ "learning_rate": 0.00010010555946516538,
1202
+ "loss": 0.356,
1203
+ "step": 3340
1204
+ },
1205
+ {
1206
+ "epoch": 1.0866752910737387,
1207
+ "grad_norm": 0.3103466331958771,
1208
+ "learning_rate": 9.940182969739622e-05,
1209
+ "loss": 0.3478,
1210
+ "step": 3360
1211
+ },
1212
+ {
1213
+ "epoch": 1.093143596377749,
1214
+ "grad_norm": 0.3601842224597931,
1215
+ "learning_rate": 9.869809992962703e-05,
1216
+ "loss": 0.3635,
1217
+ "step": 3380
1218
+ },
1219
+ {
1220
+ "epoch": 1.0996119016817594,
1221
+ "grad_norm": 0.3715918958187103,
1222
+ "learning_rate": 9.799437016185785e-05,
1223
+ "loss": 0.3489,
1224
+ "step": 3400
1225
+ },
1226
+ {
1227
+ "epoch": 1.1060802069857698,
1228
+ "grad_norm": 0.4216213524341583,
1229
+ "learning_rate": 9.729064039408867e-05,
1230
+ "loss": 0.3461,
1231
+ "step": 3420
1232
+ },
1233
+ {
1234
+ "epoch": 1.11254851228978,
1235
+ "grad_norm": 0.3878704905509949,
1236
+ "learning_rate": 9.65869106263195e-05,
1237
+ "loss": 0.3548,
1238
+ "step": 3440
1239
+ },
1240
+ {
1241
+ "epoch": 1.1190168175937905,
1242
+ "grad_norm": 0.36658766865730286,
1243
+ "learning_rate": 9.588318085855032e-05,
1244
+ "loss": 0.3395,
1245
+ "step": 3460
1246
+ },
1247
+ {
1248
+ "epoch": 1.1254851228978007,
1249
+ "grad_norm": 0.38134855031967163,
1250
+ "learning_rate": 9.517945109078115e-05,
1251
+ "loss": 0.3829,
1252
+ "step": 3480
1253
+ },
1254
+ {
1255
+ "epoch": 1.131953428201811,
1256
+ "grad_norm": 0.42683905363082886,
1257
+ "learning_rate": 9.447572132301197e-05,
1258
+ "loss": 0.3914,
1259
+ "step": 3500
1260
+ },
1261
+ {
1262
+ "epoch": 1.1384217335058215,
1263
+ "grad_norm": 0.3340086042881012,
1264
+ "learning_rate": 9.377199155524279e-05,
1265
+ "loss": 0.3301,
1266
+ "step": 3520
1267
+ },
1268
+ {
1269
+ "epoch": 1.1448900388098318,
1270
+ "grad_norm": 0.37897104024887085,
1271
+ "learning_rate": 9.306826178747361e-05,
1272
+ "loss": 0.3536,
1273
+ "step": 3540
1274
+ },
1275
+ {
1276
+ "epoch": 1.1513583441138422,
1277
+ "grad_norm": 0.361061692237854,
1278
+ "learning_rate": 9.236453201970444e-05,
1279
+ "loss": 0.338,
1280
+ "step": 3560
1281
+ },
1282
+ {
1283
+ "epoch": 1.1578266494178524,
1284
+ "grad_norm": 0.34730997681617737,
1285
+ "learning_rate": 9.166080225193527e-05,
1286
+ "loss": 0.3679,
1287
+ "step": 3580
1288
+ },
1289
+ {
1290
+ "epoch": 1.1642949547218628,
1291
+ "grad_norm": 0.29455122351646423,
1292
+ "learning_rate": 9.095707248416608e-05,
1293
+ "loss": 0.3602,
1294
+ "step": 3600
1295
+ },
1296
+ {
1297
+ "epoch": 1.1707632600258733,
1298
+ "grad_norm": 0.33159056305885315,
1299
+ "learning_rate": 9.02533427163969e-05,
1300
+ "loss": 0.3359,
1301
+ "step": 3620
1302
+ },
1303
+ {
1304
+ "epoch": 1.1772315653298835,
1305
+ "grad_norm": 0.48726096749305725,
1306
+ "learning_rate": 8.954961294862773e-05,
1307
+ "loss": 0.3879,
1308
+ "step": 3640
1309
+ },
1310
+ {
1311
+ "epoch": 1.183699870633894,
1312
+ "grad_norm": 0.3182481825351715,
1313
+ "learning_rate": 8.884588318085856e-05,
1314
+ "loss": 0.3392,
1315
+ "step": 3660
1316
+ },
1317
+ {
1318
+ "epoch": 1.1901681759379044,
1319
+ "grad_norm": 0.3488766551017761,
1320
+ "learning_rate": 8.814215341308939e-05,
1321
+ "loss": 0.3419,
1322
+ "step": 3680
1323
+ },
1324
+ {
1325
+ "epoch": 1.1966364812419146,
1326
+ "grad_norm": 0.34345540404319763,
1327
+ "learning_rate": 8.74384236453202e-05,
1328
+ "loss": 0.3598,
1329
+ "step": 3700
1330
+ },
1331
+ {
1332
+ "epoch": 1.203104786545925,
1333
+ "grad_norm": 0.37642374634742737,
1334
+ "learning_rate": 8.673469387755102e-05,
1335
+ "loss": 0.3578,
1336
+ "step": 3720
1337
+ },
1338
+ {
1339
+ "epoch": 1.2095730918499352,
1340
+ "grad_norm": 0.3726005554199219,
1341
+ "learning_rate": 8.603096410978185e-05,
1342
+ "loss": 0.3547,
1343
+ "step": 3740
1344
+ },
1345
+ {
1346
+ "epoch": 1.2160413971539457,
1347
+ "grad_norm": 0.36603832244873047,
1348
+ "learning_rate": 8.532723434201268e-05,
1349
+ "loss": 0.3831,
1350
+ "step": 3760
1351
+ },
1352
+ {
1353
+ "epoch": 1.222509702457956,
1354
+ "grad_norm": 0.345103919506073,
1355
+ "learning_rate": 8.462350457424349e-05,
1356
+ "loss": 0.3563,
1357
+ "step": 3780
1358
+ },
1359
+ {
1360
+ "epoch": 1.2289780077619663,
1361
+ "grad_norm": 0.31071504950523376,
1362
+ "learning_rate": 8.391977480647432e-05,
1363
+ "loss": 0.3263,
1364
+ "step": 3800
1365
+ },
1366
+ {
1367
+ "epoch": 1.2354463130659767,
1368
+ "grad_norm": 0.3534263074398041,
1369
+ "learning_rate": 8.321604503870513e-05,
1370
+ "loss": 0.3591,
1371
+ "step": 3820
1372
+ },
1373
+ {
1374
+ "epoch": 1.2419146183699872,
1375
+ "grad_norm": 0.3568720817565918,
1376
+ "learning_rate": 8.251231527093597e-05,
1377
+ "loss": 0.3815,
1378
+ "step": 3840
1379
+ },
1380
+ {
1381
+ "epoch": 1.2483829236739974,
1382
+ "grad_norm": 0.3450963795185089,
1383
+ "learning_rate": 8.18085855031668e-05,
1384
+ "loss": 0.3481,
1385
+ "step": 3860
1386
+ },
1387
+ {
1388
+ "epoch": 1.2548512289780078,
1389
+ "grad_norm": 0.3554648160934448,
1390
+ "learning_rate": 8.110485573539761e-05,
1391
+ "loss": 0.3364,
1392
+ "step": 3880
1393
+ },
1394
+ {
1395
+ "epoch": 1.261319534282018,
1396
+ "grad_norm": 0.4188801348209381,
1397
+ "learning_rate": 8.040112596762844e-05,
1398
+ "loss": 0.3686,
1399
+ "step": 3900
1400
+ },
1401
+ {
1402
+ "epoch": 1.2677878395860285,
1403
+ "grad_norm": 0.3129253089427948,
1404
+ "learning_rate": 7.969739619985925e-05,
1405
+ "loss": 0.3664,
1406
+ "step": 3920
1407
+ },
1408
+ {
1409
+ "epoch": 1.274256144890039,
1410
+ "grad_norm": 0.3885388970375061,
1411
+ "learning_rate": 7.899366643209009e-05,
1412
+ "loss": 0.3415,
1413
+ "step": 3940
1414
+ },
1415
+ {
1416
+ "epoch": 1.2807244501940491,
1417
+ "grad_norm": 0.4131453335285187,
1418
+ "learning_rate": 7.82899366643209e-05,
1419
+ "loss": 0.3552,
1420
+ "step": 3960
1421
+ },
1422
+ {
1423
+ "epoch": 1.2871927554980596,
1424
+ "grad_norm": 0.39486396312713623,
1425
+ "learning_rate": 7.758620689655173e-05,
1426
+ "loss": 0.366,
1427
+ "step": 3980
1428
+ },
1429
+ {
1430
+ "epoch": 1.2936610608020698,
1431
+ "grad_norm": 0.3580794930458069,
1432
+ "learning_rate": 7.688247712878254e-05,
1433
+ "loss": 0.3585,
1434
+ "step": 4000
1435
+ },
1436
+ {
1437
+ "epoch": 1.2936610608020698,
1438
+ "eval_loss": 0.5952740907669067,
1439
+ "eval_runtime": 967.8131,
1440
+ "eval_samples_per_second": 6.331,
1441
+ "eval_steps_per_second": 0.791,
1442
+ "step": 4000
1443
+ },
1444
+ {
1445
+ "epoch": 1.3001293661060802,
1446
+ "grad_norm": 0.3903474509716034,
1447
+ "learning_rate": 7.617874736101337e-05,
1448
+ "loss": 0.3347,
1449
+ "step": 4020
1450
+ },
1451
+ {
1452
+ "epoch": 1.3065976714100906,
1453
+ "grad_norm": 0.4154898226261139,
1454
+ "learning_rate": 7.54750175932442e-05,
1455
+ "loss": 0.3314,
1456
+ "step": 4040
1457
+ },
1458
+ {
1459
+ "epoch": 1.3130659767141009,
1460
+ "grad_norm": 0.33763521909713745,
1461
+ "learning_rate": 7.477128782547502e-05,
1462
+ "loss": 0.3552,
1463
+ "step": 4060
1464
+ },
1465
+ {
1466
+ "epoch": 1.3195342820181113,
1467
+ "grad_norm": 0.5618635416030884,
1468
+ "learning_rate": 7.406755805770585e-05,
1469
+ "loss": 0.363,
1470
+ "step": 4080
1471
+ },
1472
+ {
1473
+ "epoch": 1.3260025873221215,
1474
+ "grad_norm": 0.37114137411117554,
1475
+ "learning_rate": 7.336382828993666e-05,
1476
+ "loss": 0.3421,
1477
+ "step": 4100
1478
+ },
1479
+ {
1480
+ "epoch": 1.332470892626132,
1481
+ "grad_norm": 0.34323641657829285,
1482
+ "learning_rate": 7.266009852216749e-05,
1483
+ "loss": 0.3659,
1484
+ "step": 4120
1485
+ },
1486
+ {
1487
+ "epoch": 1.3389391979301424,
1488
+ "grad_norm": 0.3527199625968933,
1489
+ "learning_rate": 7.195636875439831e-05,
1490
+ "loss": 0.3412,
1491
+ "step": 4140
1492
+ },
1493
+ {
1494
+ "epoch": 1.3454075032341526,
1495
+ "grad_norm": 0.3413406014442444,
1496
+ "learning_rate": 7.125263898662914e-05,
1497
+ "loss": 0.338,
1498
+ "step": 4160
1499
+ },
1500
+ {
1501
+ "epoch": 1.351875808538163,
1502
+ "grad_norm": 0.34787651896476746,
1503
+ "learning_rate": 7.054890921885995e-05,
1504
+ "loss": 0.3681,
1505
+ "step": 4180
1506
+ },
1507
+ {
1508
+ "epoch": 1.3583441138421732,
1509
+ "grad_norm": 0.2709845006465912,
1510
+ "learning_rate": 6.984517945109078e-05,
1511
+ "loss": 0.325,
1512
+ "step": 4200
1513
+ },
1514
+ {
1515
+ "epoch": 1.3648124191461837,
1516
+ "grad_norm": 0.30966514348983765,
1517
+ "learning_rate": 6.91414496833216e-05,
1518
+ "loss": 0.3677,
1519
+ "step": 4220
1520
+ },
1521
+ {
1522
+ "epoch": 1.371280724450194,
1523
+ "grad_norm": 0.35980409383773804,
1524
+ "learning_rate": 6.843771991555243e-05,
1525
+ "loss": 0.335,
1526
+ "step": 4240
1527
+ },
1528
+ {
1529
+ "epoch": 1.3777490297542043,
1530
+ "grad_norm": 0.5173650979995728,
1531
+ "learning_rate": 6.773399014778326e-05,
1532
+ "loss": 0.348,
1533
+ "step": 4260
1534
+ },
1535
+ {
1536
+ "epoch": 1.3842173350582148,
1537
+ "grad_norm": 0.3048713207244873,
1538
+ "learning_rate": 6.703026038001407e-05,
1539
+ "loss": 0.3323,
1540
+ "step": 4280
1541
+ },
1542
+ {
1543
+ "epoch": 1.390685640362225,
1544
+ "grad_norm": 0.3427985906600952,
1545
+ "learning_rate": 6.63265306122449e-05,
1546
+ "loss": 0.3571,
1547
+ "step": 4300
1548
+ },
1549
+ {
1550
+ "epoch": 1.3971539456662354,
1551
+ "grad_norm": 0.406012624502182,
1552
+ "learning_rate": 6.562280084447573e-05,
1553
+ "loss": 0.3546,
1554
+ "step": 4320
1555
+ },
1556
+ {
1557
+ "epoch": 1.4036222509702458,
1558
+ "grad_norm": 0.31981325149536133,
1559
+ "learning_rate": 6.491907107670655e-05,
1560
+ "loss": 0.3281,
1561
+ "step": 4340
1562
+ },
1563
+ {
1564
+ "epoch": 1.4100905562742563,
1565
+ "grad_norm": 0.3936704099178314,
1566
+ "learning_rate": 6.421534130893737e-05,
1567
+ "loss": 0.3444,
1568
+ "step": 4360
1569
+ },
1570
+ {
1571
+ "epoch": 1.4165588615782665,
1572
+ "grad_norm": 0.3658936023712158,
1573
+ "learning_rate": 6.351161154116819e-05,
1574
+ "loss": 0.339,
1575
+ "step": 4380
1576
+ },
1577
+ {
1578
+ "epoch": 1.4230271668822767,
1579
+ "grad_norm": 0.32523271441459656,
1580
+ "learning_rate": 6.280788177339902e-05,
1581
+ "loss": 0.3311,
1582
+ "step": 4400
1583
+ },
1584
+ {
1585
+ "epoch": 1.4294954721862871,
1586
+ "grad_norm": 0.4120417833328247,
1587
+ "learning_rate": 6.210415200562984e-05,
1588
+ "loss": 0.3342,
1589
+ "step": 4420
1590
+ },
1591
+ {
1592
+ "epoch": 1.4359637774902976,
1593
+ "grad_norm": 0.35293978452682495,
1594
+ "learning_rate": 6.140042223786067e-05,
1595
+ "loss": 0.3561,
1596
+ "step": 4440
1597
+ },
1598
+ {
1599
+ "epoch": 1.442432082794308,
1600
+ "grad_norm": 0.33231040835380554,
1601
+ "learning_rate": 6.0696692470091484e-05,
1602
+ "loss": 0.3573,
1603
+ "step": 4460
1604
+ },
1605
+ {
1606
+ "epoch": 1.4489003880983182,
1607
+ "grad_norm": 0.4345926344394684,
1608
+ "learning_rate": 5.999296270232231e-05,
1609
+ "loss": 0.3364,
1610
+ "step": 4480
1611
+ },
1612
+ {
1613
+ "epoch": 1.4553686934023287,
1614
+ "grad_norm": 0.34958288073539734,
1615
+ "learning_rate": 5.928923293455313e-05,
1616
+ "loss": 0.3162,
1617
+ "step": 4500
1618
+ },
1619
+ {
1620
+ "epoch": 1.4618369987063389,
1621
+ "grad_norm": 0.4669477045536041,
1622
+ "learning_rate": 5.8585503166783964e-05,
1623
+ "loss": 0.3455,
1624
+ "step": 4520
1625
+ },
1626
+ {
1627
+ "epoch": 1.4683053040103493,
1628
+ "grad_norm": 0.32660993933677673,
1629
+ "learning_rate": 5.788177339901478e-05,
1630
+ "loss": 0.3228,
1631
+ "step": 4540
1632
+ },
1633
+ {
1634
+ "epoch": 1.4747736093143597,
1635
+ "grad_norm": 0.3654051423072815,
1636
+ "learning_rate": 5.71780436312456e-05,
1637
+ "loss": 0.3293,
1638
+ "step": 4560
1639
+ },
1640
+ {
1641
+ "epoch": 1.48124191461837,
1642
+ "grad_norm": 0.3496916592121124,
1643
+ "learning_rate": 5.647431386347642e-05,
1644
+ "loss": 0.3382,
1645
+ "step": 4580
1646
+ },
1647
+ {
1648
+ "epoch": 1.4877102199223804,
1649
+ "grad_norm": 0.3890388309955597,
1650
+ "learning_rate": 5.577058409570724e-05,
1651
+ "loss": 0.3468,
1652
+ "step": 4600
1653
+ },
1654
+ {
1655
+ "epoch": 1.4941785252263906,
1656
+ "grad_norm": 0.3924092650413513,
1657
+ "learning_rate": 5.5066854327938076e-05,
1658
+ "loss": 0.3467,
1659
+ "step": 4620
1660
+ },
1661
+ {
1662
+ "epoch": 1.500646830530401,
1663
+ "grad_norm": 0.4015033543109894,
1664
+ "learning_rate": 5.4363124560168896e-05,
1665
+ "loss": 0.3478,
1666
+ "step": 4640
1667
+ },
1668
+ {
1669
+ "epoch": 1.5071151358344115,
1670
+ "grad_norm": 0.34742093086242676,
1671
+ "learning_rate": 5.365939479239972e-05,
1672
+ "loss": 0.3252,
1673
+ "step": 4660
1674
+ },
1675
+ {
1676
+ "epoch": 1.5135834411384217,
1677
+ "grad_norm": 0.3477989435195923,
1678
+ "learning_rate": 5.295566502463054e-05,
1679
+ "loss": 0.3627,
1680
+ "step": 4680
1681
+ },
1682
+ {
1683
+ "epoch": 1.5200517464424321,
1684
+ "grad_norm": 0.4194832146167755,
1685
+ "learning_rate": 5.2251935256861375e-05,
1686
+ "loss": 0.3403,
1687
+ "step": 4700
1688
+ },
1689
+ {
1690
+ "epoch": 1.5265200517464423,
1691
+ "grad_norm": 0.37723594903945923,
1692
+ "learning_rate": 5.1548205489092195e-05,
1693
+ "loss": 0.3424,
1694
+ "step": 4720
1695
+ },
1696
+ {
1697
+ "epoch": 1.5329883570504528,
1698
+ "grad_norm": 0.3695043921470642,
1699
+ "learning_rate": 5.0844475721323015e-05,
1700
+ "loss": 0.3339,
1701
+ "step": 4740
1702
+ },
1703
+ {
1704
+ "epoch": 1.5394566623544632,
1705
+ "grad_norm": 0.35015031695365906,
1706
+ "learning_rate": 5.0140745953553834e-05,
1707
+ "loss": 0.3262,
1708
+ "step": 4760
1709
+ },
1710
+ {
1711
+ "epoch": 1.5459249676584734,
1712
+ "grad_norm": 0.39946597814559937,
1713
+ "learning_rate": 4.943701618578466e-05,
1714
+ "loss": 0.3649,
1715
+ "step": 4780
1716
+ },
1717
+ {
1718
+ "epoch": 1.5523932729624839,
1719
+ "grad_norm": 0.3246670961380005,
1720
+ "learning_rate": 4.873328641801548e-05,
1721
+ "loss": 0.3241,
1722
+ "step": 4800
1723
+ },
1724
+ {
1725
+ "epoch": 1.558861578266494,
1726
+ "grad_norm": 0.41913464665412903,
1727
+ "learning_rate": 4.802955665024631e-05,
1728
+ "loss": 0.3486,
1729
+ "step": 4820
1730
+ },
1731
+ {
1732
+ "epoch": 1.5653298835705045,
1733
+ "grad_norm": 0.38162580132484436,
1734
+ "learning_rate": 4.7325826882477134e-05,
1735
+ "loss": 0.3571,
1736
+ "step": 4840
1737
+ },
1738
+ {
1739
+ "epoch": 1.571798188874515,
1740
+ "grad_norm": 0.34426695108413696,
1741
+ "learning_rate": 4.6622097114707953e-05,
1742
+ "loss": 0.3232,
1743
+ "step": 4860
1744
+ },
1745
+ {
1746
+ "epoch": 1.5782664941785254,
1747
+ "grad_norm": 0.3782813847064972,
1748
+ "learning_rate": 4.591836734693878e-05,
1749
+ "loss": 0.346,
1750
+ "step": 4880
1751
+ },
1752
+ {
1753
+ "epoch": 1.5847347994825356,
1754
+ "grad_norm": 0.42736709117889404,
1755
+ "learning_rate": 4.52146375791696e-05,
1756
+ "loss": 0.3275,
1757
+ "step": 4900
1758
+ },
1759
+ {
1760
+ "epoch": 1.5912031047865458,
1761
+ "grad_norm": 0.37077978253364563,
1762
+ "learning_rate": 4.4510907811400426e-05,
1763
+ "loss": 0.3312,
1764
+ "step": 4920
1765
+ },
1766
+ {
1767
+ "epoch": 1.5976714100905562,
1768
+ "grad_norm": 0.3196958303451538,
1769
+ "learning_rate": 4.3807178043631246e-05,
1770
+ "loss": 0.3156,
1771
+ "step": 4940
1772
+ },
1773
+ {
1774
+ "epoch": 1.6041397153945667,
1775
+ "grad_norm": 0.36680254340171814,
1776
+ "learning_rate": 4.3103448275862066e-05,
1777
+ "loss": 0.3293,
1778
+ "step": 4960
1779
+ },
1780
+ {
1781
+ "epoch": 1.6106080206985771,
1782
+ "grad_norm": 0.44455257058143616,
1783
+ "learning_rate": 4.239971850809289e-05,
1784
+ "loss": 0.3512,
1785
+ "step": 4980
1786
+ },
1787
+ {
1788
+ "epoch": 1.6170763260025873,
1789
+ "grad_norm": 0.37418290972709656,
1790
+ "learning_rate": 4.169598874032372e-05,
1791
+ "loss": 0.3422,
1792
+ "step": 5000
1793
+ },
1794
+ {
1795
+ "epoch": 1.6170763260025873,
1796
+ "eval_loss": 0.5881822109222412,
1797
+ "eval_runtime": 967.8912,
1798
+ "eval_samples_per_second": 6.33,
1799
+ "eval_steps_per_second": 0.791,
1800
+ "step": 5000
1801
+ }
1802
+ ],
1803
+ "logging_steps": 20,
1804
+ "max_steps": 6184,
1805
+ "num_input_tokens_seen": 0,
1806
+ "num_train_epochs": 2,
1807
+ "save_steps": 1000,
1808
+ "stateful_callbacks": {
1809
+ "TrainerControl": {
1810
+ "args": {
1811
+ "should_epoch_stop": false,
1812
+ "should_evaluate": false,
1813
+ "should_log": false,
1814
+ "should_save": true,
1815
+ "should_training_stop": false
1816
+ },
1817
+ "attributes": {}
1818
+ }
1819
+ },
1820
+ "total_flos": 6.669845911961469e+18,
1821
+ "train_batch_size": 8,
1822
+ "trial_name": null,
1823
+ "trial_params": null
1824
+ }
ultival_mistral/checkpoint-5000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8e5db938f70ff2ddfe787ade649d14bb55cc2be7a65bd63a20c3303bf68492
3
+ size 5368
ultival_mistral/checkpoint-6000/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Ministral-8B-Instruct-2410
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
ultival_mistral/checkpoint-6000/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Ministral-8B-Instruct-2410",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 32,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "o_proj",
28
+ "k_proj",
29
+ "down_proj",
30
+ "q_proj",
31
+ "gate_proj",
32
+ "up_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
ultival_mistral/checkpoint-6000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f082f6d0a3a97b05923cb2b2b2fbe02ce9f81b4b6db500a43193fc1828f6cdff
3
+ size 349243752
ultival_mistral/checkpoint-6000/chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{% if messages[1]['role'] == 'user' %}{{ '[INST] ' + messages[0]['content'] + ' ' + messages[1]['content'] + ' [/INST]' }}{% set loop_messages = messages[2:] %}{% else %}{{ '[INST] ' + messages[0]['content'] + ' [/INST]' }}{% set loop_messages = messages[1:] %}{% endif %}{% else %}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}
ultival_mistral/checkpoint-6000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f55137f8a9b9997a59aaeb82cd818e104ce757097a31612eb87abed422dce57b
3
+ size 177919028
ultival_mistral/checkpoint-6000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf932362a907148c2c99f7826e21fa7280b5016d990a6358e60ec3cec98b016
3
+ size 14244
ultival_mistral/checkpoint-6000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ed7383692d087a8f053c6ea044fdaeeeda6d990035145da8af603021621b27
3
+ size 1064
ultival_mistral/checkpoint-6000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<pad>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
ultival_mistral/checkpoint-6000/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be2965d3e435dbdef9376e5bdb3dcc2bfc7e271aa5152ae896cf53c7e920adf8
3
+ size 17078235
ultival_mistral/checkpoint-6000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
ultival_mistral/checkpoint-6000/trainer_state.json ADDED
@@ -0,0 +1,2182 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 6000,
3
+ "best_metric": 0.5839738845825195,
4
+ "best_model_checkpoint": "./ultival_run2/checkpoint-6000",
5
+ "epoch": 1.9404915912031049,
6
+ "eval_steps": 1000,
7
+ "global_step": 6000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00646830530401035,
14
+ "grad_norm": 0.8976479172706604,
15
+ "learning_rate": 7.6e-06,
16
+ "loss": 1.5544,
17
+ "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.0129366106080207,
21
+ "grad_norm": 0.726784348487854,
22
+ "learning_rate": 1.56e-05,
23
+ "loss": 1.242,
24
+ "step": 40
25
+ },
26
+ {
27
+ "epoch": 0.019404915912031046,
28
+ "grad_norm": 0.4917074143886566,
29
+ "learning_rate": 2.36e-05,
30
+ "loss": 0.6694,
31
+ "step": 60
32
+ },
33
+ {
34
+ "epoch": 0.0258732212160414,
35
+ "grad_norm": 0.3471176326274872,
36
+ "learning_rate": 3.16e-05,
37
+ "loss": 0.5334,
38
+ "step": 80
39
+ },
40
+ {
41
+ "epoch": 0.03234152652005175,
42
+ "grad_norm": 0.41759616136550903,
43
+ "learning_rate": 3.960000000000001e-05,
44
+ "loss": 0.5435,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.03880983182406209,
49
+ "grad_norm": 0.3755040466785431,
50
+ "learning_rate": 4.76e-05,
51
+ "loss": 0.5684,
52
+ "step": 120
53
+ },
54
+ {
55
+ "epoch": 0.045278137128072445,
56
+ "grad_norm": 0.37058988213539124,
57
+ "learning_rate": 5.560000000000001e-05,
58
+ "loss": 0.5329,
59
+ "step": 140
60
+ },
61
+ {
62
+ "epoch": 0.0517464424320828,
63
+ "grad_norm": 0.42603132128715515,
64
+ "learning_rate": 6.36e-05,
65
+ "loss": 0.5457,
66
+ "step": 160
67
+ },
68
+ {
69
+ "epoch": 0.05821474773609314,
70
+ "grad_norm": 0.3154866099357605,
71
+ "learning_rate": 7.16e-05,
72
+ "loss": 0.5472,
73
+ "step": 180
74
+ },
75
+ {
76
+ "epoch": 0.0646830530401035,
77
+ "grad_norm": 0.3919835388660431,
78
+ "learning_rate": 7.960000000000001e-05,
79
+ "loss": 0.4675,
80
+ "step": 200
81
+ },
82
+ {
83
+ "epoch": 0.07115135834411385,
84
+ "grad_norm": 0.32180657982826233,
85
+ "learning_rate": 8.76e-05,
86
+ "loss": 0.5078,
87
+ "step": 220
88
+ },
89
+ {
90
+ "epoch": 0.07761966364812418,
91
+ "grad_norm": 0.3246982991695404,
92
+ "learning_rate": 9.56e-05,
93
+ "loss": 0.488,
94
+ "step": 240
95
+ },
96
+ {
97
+ "epoch": 0.08408796895213454,
98
+ "grad_norm": 0.341487318277359,
99
+ "learning_rate": 0.00010360000000000001,
100
+ "loss": 0.5058,
101
+ "step": 260
102
+ },
103
+ {
104
+ "epoch": 0.09055627425614489,
105
+ "grad_norm": 0.3597952127456665,
106
+ "learning_rate": 0.00011160000000000002,
107
+ "loss": 0.4574,
108
+ "step": 280
109
+ },
110
+ {
111
+ "epoch": 0.09702457956015524,
112
+ "grad_norm": 0.40912091732025146,
113
+ "learning_rate": 0.00011960000000000001,
114
+ "loss": 0.4931,
115
+ "step": 300
116
+ },
117
+ {
118
+ "epoch": 0.1034928848641656,
119
+ "grad_norm": 0.37430697679519653,
120
+ "learning_rate": 0.0001276,
121
+ "loss": 0.4724,
122
+ "step": 320
123
+ },
124
+ {
125
+ "epoch": 0.10996119016817593,
126
+ "grad_norm": 0.3920972943305969,
127
+ "learning_rate": 0.00013560000000000002,
128
+ "loss": 0.4997,
129
+ "step": 340
130
+ },
131
+ {
132
+ "epoch": 0.11642949547218628,
133
+ "grad_norm": 0.32511982321739197,
134
+ "learning_rate": 0.0001436,
135
+ "loss": 0.4889,
136
+ "step": 360
137
+ },
138
+ {
139
+ "epoch": 0.12289780077619664,
140
+ "grad_norm": 0.3926583528518677,
141
+ "learning_rate": 0.0001516,
142
+ "loss": 0.4772,
143
+ "step": 380
144
+ },
145
+ {
146
+ "epoch": 0.129366106080207,
147
+ "grad_norm": 0.3535909950733185,
148
+ "learning_rate": 0.0001596,
149
+ "loss": 0.465,
150
+ "step": 400
151
+ },
152
+ {
153
+ "epoch": 0.13583441138421734,
154
+ "grad_norm": 0.34381625056266785,
155
+ "learning_rate": 0.0001676,
156
+ "loss": 0.4934,
157
+ "step": 420
158
+ },
159
+ {
160
+ "epoch": 0.1423027166882277,
161
+ "grad_norm": 0.3753703236579895,
162
+ "learning_rate": 0.0001756,
163
+ "loss": 0.5068,
164
+ "step": 440
165
+ },
166
+ {
167
+ "epoch": 0.14877102199223805,
168
+ "grad_norm": 0.34895479679107666,
169
+ "learning_rate": 0.00018360000000000002,
170
+ "loss": 0.5186,
171
+ "step": 460
172
+ },
173
+ {
174
+ "epoch": 0.15523932729624837,
175
+ "grad_norm": 0.24720332026481628,
176
+ "learning_rate": 0.0001916,
177
+ "loss": 0.461,
178
+ "step": 480
179
+ },
180
+ {
181
+ "epoch": 0.16170763260025872,
182
+ "grad_norm": 0.3713696002960205,
183
+ "learning_rate": 0.0001996,
184
+ "loss": 0.4932,
185
+ "step": 500
186
+ },
187
+ {
188
+ "epoch": 0.16817593790426907,
189
+ "grad_norm": 0.3724994659423828,
190
+ "learning_rate": 0.00019933145672061927,
191
+ "loss": 0.5036,
192
+ "step": 520
193
+ },
194
+ {
195
+ "epoch": 0.17464424320827943,
196
+ "grad_norm": 0.3562956750392914,
197
+ "learning_rate": 0.00019862772695285012,
198
+ "loss": 0.4871,
199
+ "step": 540
200
+ },
201
+ {
202
+ "epoch": 0.18111254851228978,
203
+ "grad_norm": 0.331737756729126,
204
+ "learning_rate": 0.00019792399718508095,
205
+ "loss": 0.5178,
206
+ "step": 560
207
+ },
208
+ {
209
+ "epoch": 0.18758085381630013,
210
+ "grad_norm": 0.34049859642982483,
211
+ "learning_rate": 0.00019722026741731178,
212
+ "loss": 0.488,
213
+ "step": 580
214
+ },
215
+ {
216
+ "epoch": 0.19404915912031048,
217
+ "grad_norm": 0.36076661944389343,
218
+ "learning_rate": 0.00019651653764954258,
219
+ "loss": 0.4821,
220
+ "step": 600
221
+ },
222
+ {
223
+ "epoch": 0.20051746442432083,
224
+ "grad_norm": 0.3059399127960205,
225
+ "learning_rate": 0.0001958128078817734,
226
+ "loss": 0.4924,
227
+ "step": 620
228
+ },
229
+ {
230
+ "epoch": 0.2069857697283312,
231
+ "grad_norm": 0.30692920088768005,
232
+ "learning_rate": 0.00019510907811400423,
233
+ "loss": 0.4822,
234
+ "step": 640
235
+ },
236
+ {
237
+ "epoch": 0.21345407503234154,
238
+ "grad_norm": 0.3319448232650757,
239
+ "learning_rate": 0.00019440534834623506,
240
+ "loss": 0.4607,
241
+ "step": 660
242
+ },
243
+ {
244
+ "epoch": 0.21992238033635186,
245
+ "grad_norm": 0.3047769069671631,
246
+ "learning_rate": 0.00019370161857846588,
247
+ "loss": 0.5357,
248
+ "step": 680
249
+ },
250
+ {
251
+ "epoch": 0.22639068564036222,
252
+ "grad_norm": 0.4353258013725281,
253
+ "learning_rate": 0.00019299788881069668,
254
+ "loss": 0.4738,
255
+ "step": 700
256
+ },
257
+ {
258
+ "epoch": 0.23285899094437257,
259
+ "grad_norm": 0.30877336859703064,
260
+ "learning_rate": 0.00019229415904292754,
261
+ "loss": 0.5128,
262
+ "step": 720
263
+ },
264
+ {
265
+ "epoch": 0.23932729624838292,
266
+ "grad_norm": 0.3188530206680298,
267
+ "learning_rate": 0.00019159042927515836,
268
+ "loss": 0.4651,
269
+ "step": 740
270
+ },
271
+ {
272
+ "epoch": 0.24579560155239327,
273
+ "grad_norm": 0.40678778290748596,
274
+ "learning_rate": 0.0001908866995073892,
275
+ "loss": 0.5112,
276
+ "step": 760
277
+ },
278
+ {
279
+ "epoch": 0.2522639068564036,
280
+ "grad_norm": 0.34188327193260193,
281
+ "learning_rate": 0.00019018296973962,
282
+ "loss": 0.487,
283
+ "step": 780
284
+ },
285
+ {
286
+ "epoch": 0.258732212160414,
287
+ "grad_norm": 0.3002854585647583,
288
+ "learning_rate": 0.00018947923997185081,
289
+ "loss": 0.512,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.2652005174644243,
294
+ "grad_norm": 0.35756272077560425,
295
+ "learning_rate": 0.00018877551020408164,
296
+ "loss": 0.4812,
297
+ "step": 820
298
+ },
299
+ {
300
+ "epoch": 0.2716688227684347,
301
+ "grad_norm": 0.33527258038520813,
302
+ "learning_rate": 0.00018807178043631247,
303
+ "loss": 0.4895,
304
+ "step": 840
305
+ },
306
+ {
307
+ "epoch": 0.278137128072445,
308
+ "grad_norm": 0.3472883105278015,
309
+ "learning_rate": 0.0001873680506685433,
310
+ "loss": 0.4671,
311
+ "step": 860
312
+ },
313
+ {
314
+ "epoch": 0.2846054333764554,
315
+ "grad_norm": 0.3604671061038971,
316
+ "learning_rate": 0.0001866643209007741,
317
+ "loss": 0.5233,
318
+ "step": 880
319
+ },
320
+ {
321
+ "epoch": 0.2910737386804657,
322
+ "grad_norm": 0.31021153926849365,
323
+ "learning_rate": 0.00018596059113300492,
324
+ "loss": 0.491,
325
+ "step": 900
326
+ },
327
+ {
328
+ "epoch": 0.2975420439844761,
329
+ "grad_norm": 0.6417067050933838,
330
+ "learning_rate": 0.00018525686136523577,
331
+ "loss": 0.4858,
332
+ "step": 920
333
+ },
334
+ {
335
+ "epoch": 0.3040103492884864,
336
+ "grad_norm": 0.33719781041145325,
337
+ "learning_rate": 0.0001845531315974666,
338
+ "loss": 0.4722,
339
+ "step": 940
340
+ },
341
+ {
342
+ "epoch": 0.31047865459249674,
343
+ "grad_norm": 0.3568169176578522,
344
+ "learning_rate": 0.0001838494018296974,
345
+ "loss": 0.5088,
346
+ "step": 960
347
+ },
348
+ {
349
+ "epoch": 0.3169469598965071,
350
+ "grad_norm": 0.42162615060806274,
351
+ "learning_rate": 0.00018314567206192823,
352
+ "loss": 0.5091,
353
+ "step": 980
354
+ },
355
+ {
356
+ "epoch": 0.32341526520051744,
357
+ "grad_norm": 0.3212260603904724,
358
+ "learning_rate": 0.00018244194229415905,
359
+ "loss": 0.4654,
360
+ "step": 1000
361
+ },
362
+ {
363
+ "epoch": 0.32341526520051744,
364
+ "eval_loss": 0.6167479753494263,
365
+ "eval_runtime": 1013.3659,
366
+ "eval_samples_per_second": 6.046,
367
+ "eval_steps_per_second": 0.756,
368
+ "step": 1000
369
+ },
370
+ {
371
+ "epoch": 0.3298835705045278,
372
+ "grad_norm": 0.3184376060962677,
373
+ "learning_rate": 0.00018173821252638988,
374
+ "loss": 0.4712,
375
+ "step": 1020
376
+ },
377
+ {
378
+ "epoch": 0.33635187580853815,
379
+ "grad_norm": 0.3114645183086395,
380
+ "learning_rate": 0.0001810344827586207,
381
+ "loss": 0.4393,
382
+ "step": 1040
383
+ },
384
+ {
385
+ "epoch": 0.3428201811125485,
386
+ "grad_norm": 0.30122488737106323,
387
+ "learning_rate": 0.0001803307529908515,
388
+ "loss": 0.4882,
389
+ "step": 1060
390
+ },
391
+ {
392
+ "epoch": 0.34928848641655885,
393
+ "grad_norm": 0.2877393066883087,
394
+ "learning_rate": 0.00017962702322308233,
395
+ "loss": 0.413,
396
+ "step": 1080
397
+ },
398
+ {
399
+ "epoch": 0.35575679172056923,
400
+ "grad_norm": 0.33227887749671936,
401
+ "learning_rate": 0.00017892329345531319,
402
+ "loss": 0.4527,
403
+ "step": 1100
404
+ },
405
+ {
406
+ "epoch": 0.36222509702457956,
407
+ "grad_norm": 0.33423128724098206,
408
+ "learning_rate": 0.000178219563687544,
409
+ "loss": 0.4604,
410
+ "step": 1120
411
+ },
412
+ {
413
+ "epoch": 0.36869340232858994,
414
+ "grad_norm": 0.36922487616539,
415
+ "learning_rate": 0.0001775158339197748,
416
+ "loss": 0.5044,
417
+ "step": 1140
418
+ },
419
+ {
420
+ "epoch": 0.37516170763260026,
421
+ "grad_norm": 0.34816262125968933,
422
+ "learning_rate": 0.00017681210415200564,
423
+ "loss": 0.4898,
424
+ "step": 1160
425
+ },
426
+ {
427
+ "epoch": 0.3816300129366106,
428
+ "grad_norm": 0.32086044549942017,
429
+ "learning_rate": 0.00017610837438423646,
430
+ "loss": 0.4647,
431
+ "step": 1180
432
+ },
433
+ {
434
+ "epoch": 0.38809831824062097,
435
+ "grad_norm": 0.3879236876964569,
436
+ "learning_rate": 0.0001754046446164673,
437
+ "loss": 0.5179,
438
+ "step": 1200
439
+ },
440
+ {
441
+ "epoch": 0.3945666235446313,
442
+ "grad_norm": 0.2899722158908844,
443
+ "learning_rate": 0.00017470091484869812,
444
+ "loss": 0.4786,
445
+ "step": 1220
446
+ },
447
+ {
448
+ "epoch": 0.40103492884864167,
449
+ "grad_norm": 0.3844458758831024,
450
+ "learning_rate": 0.00017399718508092892,
451
+ "loss": 0.4513,
452
+ "step": 1240
453
+ },
454
+ {
455
+ "epoch": 0.407503234152652,
456
+ "grad_norm": 0.3678169548511505,
457
+ "learning_rate": 0.00017329345531315974,
458
+ "loss": 0.4789,
459
+ "step": 1260
460
+ },
461
+ {
462
+ "epoch": 0.4139715394566624,
463
+ "grad_norm": 0.37138116359710693,
464
+ "learning_rate": 0.00017258972554539057,
465
+ "loss": 0.4694,
466
+ "step": 1280
467
+ },
468
+ {
469
+ "epoch": 0.4204398447606727,
470
+ "grad_norm": 0.3604221045970917,
471
+ "learning_rate": 0.00017188599577762142,
472
+ "loss": 0.4496,
473
+ "step": 1300
474
+ },
475
+ {
476
+ "epoch": 0.4269081500646831,
477
+ "grad_norm": 0.33961811661720276,
478
+ "learning_rate": 0.00017118226600985222,
479
+ "loss": 0.4696,
480
+ "step": 1320
481
+ },
482
+ {
483
+ "epoch": 0.4333764553686934,
484
+ "grad_norm": 0.3244430124759674,
485
+ "learning_rate": 0.00017047853624208305,
486
+ "loss": 0.4322,
487
+ "step": 1340
488
+ },
489
+ {
490
+ "epoch": 0.4398447606727037,
491
+ "grad_norm": 0.2879267930984497,
492
+ "learning_rate": 0.00016977480647431388,
493
+ "loss": 0.4646,
494
+ "step": 1360
495
+ },
496
+ {
497
+ "epoch": 0.4463130659767141,
498
+ "grad_norm": 0.3437345027923584,
499
+ "learning_rate": 0.0001690710767065447,
500
+ "loss": 0.4711,
501
+ "step": 1380
502
+ },
503
+ {
504
+ "epoch": 0.45278137128072443,
505
+ "grad_norm": 0.3166882395744324,
506
+ "learning_rate": 0.00016836734693877553,
507
+ "loss": 0.486,
508
+ "step": 1400
509
+ },
510
+ {
511
+ "epoch": 0.4592496765847348,
512
+ "grad_norm": 0.3677555322647095,
513
+ "learning_rate": 0.00016766361717100633,
514
+ "loss": 0.4677,
515
+ "step": 1420
516
+ },
517
+ {
518
+ "epoch": 0.46571798188874514,
519
+ "grad_norm": 0.32657769322395325,
520
+ "learning_rate": 0.00016695988740323715,
521
+ "loss": 0.4649,
522
+ "step": 1440
523
+ },
524
+ {
525
+ "epoch": 0.4721862871927555,
526
+ "grad_norm": 0.2980731427669525,
527
+ "learning_rate": 0.00016625615763546798,
528
+ "loss": 0.4547,
529
+ "step": 1460
530
+ },
531
+ {
532
+ "epoch": 0.47865459249676584,
533
+ "grad_norm": 0.3690549433231354,
534
+ "learning_rate": 0.0001655524278676988,
535
+ "loss": 0.4512,
536
+ "step": 1480
537
+ },
538
+ {
539
+ "epoch": 0.4851228978007762,
540
+ "grad_norm": 0.31456947326660156,
541
+ "learning_rate": 0.00016484869809992963,
542
+ "loss": 0.4675,
543
+ "step": 1500
544
+ },
545
+ {
546
+ "epoch": 0.49159120310478654,
547
+ "grad_norm": 0.3823802173137665,
548
+ "learning_rate": 0.00016414496833216046,
549
+ "loss": 0.4849,
550
+ "step": 1520
551
+ },
552
+ {
553
+ "epoch": 0.49805950840879687,
554
+ "grad_norm": 0.33280324935913086,
555
+ "learning_rate": 0.0001634412385643913,
556
+ "loss": 0.4876,
557
+ "step": 1540
558
+ },
559
+ {
560
+ "epoch": 0.5045278137128072,
561
+ "grad_norm": 0.28788062930107117,
562
+ "learning_rate": 0.00016273750879662211,
563
+ "loss": 0.426,
564
+ "step": 1560
565
+ },
566
+ {
567
+ "epoch": 0.5109961190168176,
568
+ "grad_norm": 0.3809375762939453,
569
+ "learning_rate": 0.00016203377902885294,
570
+ "loss": 0.4676,
571
+ "step": 1580
572
+ },
573
+ {
574
+ "epoch": 0.517464424320828,
575
+ "grad_norm": 0.27940312027931213,
576
+ "learning_rate": 0.00016133004926108374,
577
+ "loss": 0.4413,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 0.5239327296248383,
582
+ "grad_norm": 0.30946919322013855,
583
+ "learning_rate": 0.00016062631949331457,
584
+ "loss": 0.4395,
585
+ "step": 1620
586
+ },
587
+ {
588
+ "epoch": 0.5304010349288486,
589
+ "grad_norm": 0.36931583285331726,
590
+ "learning_rate": 0.0001599225897255454,
591
+ "loss": 0.4565,
592
+ "step": 1640
593
+ },
594
+ {
595
+ "epoch": 0.536869340232859,
596
+ "grad_norm": 0.3652470111846924,
597
+ "learning_rate": 0.00015921885995777622,
598
+ "loss": 0.4522,
599
+ "step": 1660
600
+ },
601
+ {
602
+ "epoch": 0.5433376455368694,
603
+ "grad_norm": 0.3060782551765442,
604
+ "learning_rate": 0.00015851513019000705,
605
+ "loss": 0.45,
606
+ "step": 1680
607
+ },
608
+ {
609
+ "epoch": 0.5498059508408797,
610
+ "grad_norm": 0.3275803029537201,
611
+ "learning_rate": 0.00015781140042223787,
612
+ "loss": 0.445,
613
+ "step": 1700
614
+ },
615
+ {
616
+ "epoch": 0.55627425614489,
617
+ "grad_norm": 0.34865352511405945,
618
+ "learning_rate": 0.0001571076706544687,
619
+ "loss": 0.4306,
620
+ "step": 1720
621
+ },
622
+ {
623
+ "epoch": 0.5627425614489003,
624
+ "grad_norm": 0.42525383830070496,
625
+ "learning_rate": 0.00015640394088669953,
626
+ "loss": 0.4544,
627
+ "step": 1740
628
+ },
629
+ {
630
+ "epoch": 0.5692108667529108,
631
+ "grad_norm": 0.32165050506591797,
632
+ "learning_rate": 0.00015570021111893035,
633
+ "loss": 0.484,
634
+ "step": 1760
635
+ },
636
+ {
637
+ "epoch": 0.5756791720569211,
638
+ "grad_norm": 0.3213440477848053,
639
+ "learning_rate": 0.00015499648135116115,
640
+ "loss": 0.4469,
641
+ "step": 1780
642
+ },
643
+ {
644
+ "epoch": 0.5821474773609314,
645
+ "grad_norm": 0.42835432291030884,
646
+ "learning_rate": 0.00015429275158339198,
647
+ "loss": 0.4592,
648
+ "step": 1800
649
+ },
650
+ {
651
+ "epoch": 0.5886157826649417,
652
+ "grad_norm": 0.3861497938632965,
653
+ "learning_rate": 0.0001535890218156228,
654
+ "loss": 0.4607,
655
+ "step": 1820
656
+ },
657
+ {
658
+ "epoch": 0.5950840879689522,
659
+ "grad_norm": 0.4156148135662079,
660
+ "learning_rate": 0.00015288529204785363,
661
+ "loss": 0.4541,
662
+ "step": 1840
663
+ },
664
+ {
665
+ "epoch": 0.6015523932729625,
666
+ "grad_norm": 0.32371824979782104,
667
+ "learning_rate": 0.00015218156228008446,
668
+ "loss": 0.4287,
669
+ "step": 1860
670
+ },
671
+ {
672
+ "epoch": 0.6080206985769728,
673
+ "grad_norm": 0.363788366317749,
674
+ "learning_rate": 0.00015147783251231528,
675
+ "loss": 0.428,
676
+ "step": 1880
677
+ },
678
+ {
679
+ "epoch": 0.6144890038809832,
680
+ "grad_norm": 0.3932839334011078,
681
+ "learning_rate": 0.0001507741027445461,
682
+ "loss": 0.4359,
683
+ "step": 1900
684
+ },
685
+ {
686
+ "epoch": 0.6209573091849935,
687
+ "grad_norm": 0.3009307086467743,
688
+ "learning_rate": 0.00015007037297677694,
689
+ "loss": 0.3976,
690
+ "step": 1920
691
+ },
692
+ {
693
+ "epoch": 0.6274256144890039,
694
+ "grad_norm": 0.3484659194946289,
695
+ "learning_rate": 0.00014936664320900776,
696
+ "loss": 0.4584,
697
+ "step": 1940
698
+ },
699
+ {
700
+ "epoch": 0.6338939197930142,
701
+ "grad_norm": 0.33728504180908203,
702
+ "learning_rate": 0.00014866291344123856,
703
+ "loss": 0.4543,
704
+ "step": 1960
705
+ },
706
+ {
707
+ "epoch": 0.6403622250970246,
708
+ "grad_norm": 0.3207024037837982,
709
+ "learning_rate": 0.0001479591836734694,
710
+ "loss": 0.4611,
711
+ "step": 1980
712
+ },
713
+ {
714
+ "epoch": 0.6468305304010349,
715
+ "grad_norm": 0.29134076833724976,
716
+ "learning_rate": 0.00014725545390570022,
717
+ "loss": 0.439,
718
+ "step": 2000
719
+ },
720
+ {
721
+ "epoch": 0.6468305304010349,
722
+ "eval_loss": 0.6003134250640869,
723
+ "eval_runtime": 968.0302,
724
+ "eval_samples_per_second": 6.329,
725
+ "eval_steps_per_second": 0.791,
726
+ "step": 2000
727
+ },
728
+ {
729
+ "epoch": 0.6532988357050453,
730
+ "grad_norm": 0.3548874258995056,
731
+ "learning_rate": 0.00014655172413793104,
732
+ "loss": 0.4548,
733
+ "step": 2020
734
+ },
735
+ {
736
+ "epoch": 0.6597671410090556,
737
+ "grad_norm": 0.3707274794578552,
738
+ "learning_rate": 0.00014584799437016187,
739
+ "loss": 0.4271,
740
+ "step": 2040
741
+ },
742
+ {
743
+ "epoch": 0.666235446313066,
744
+ "grad_norm": 0.37172695994377136,
745
+ "learning_rate": 0.0001451442646023927,
746
+ "loss": 0.4626,
747
+ "step": 2060
748
+ },
749
+ {
750
+ "epoch": 0.6727037516170763,
751
+ "grad_norm": 0.3459303081035614,
752
+ "learning_rate": 0.00014444053483462352,
753
+ "loss": 0.4317,
754
+ "step": 2080
755
+ },
756
+ {
757
+ "epoch": 0.6791720569210866,
758
+ "grad_norm": 0.3477213680744171,
759
+ "learning_rate": 0.00014373680506685435,
760
+ "loss": 0.4243,
761
+ "step": 2100
762
+ },
763
+ {
764
+ "epoch": 0.685640362225097,
765
+ "grad_norm": 0.31847599148750305,
766
+ "learning_rate": 0.00014303307529908517,
767
+ "loss": 0.4234,
768
+ "step": 2120
769
+ },
770
+ {
771
+ "epoch": 0.6921086675291074,
772
+ "grad_norm": 0.3231470584869385,
773
+ "learning_rate": 0.00014232934553131597,
774
+ "loss": 0.4302,
775
+ "step": 2140
776
+ },
777
+ {
778
+ "epoch": 0.6985769728331177,
779
+ "grad_norm": 0.3745759427547455,
780
+ "learning_rate": 0.0001416256157635468,
781
+ "loss": 0.4533,
782
+ "step": 2160
783
+ },
784
+ {
785
+ "epoch": 0.705045278137128,
786
+ "grad_norm": 0.34834805130958557,
787
+ "learning_rate": 0.00014092188599577763,
788
+ "loss": 0.4748,
789
+ "step": 2180
790
+ },
791
+ {
792
+ "epoch": 0.7115135834411385,
793
+ "grad_norm": 0.3675827085971832,
794
+ "learning_rate": 0.00014021815622800845,
795
+ "loss": 0.4401,
796
+ "step": 2200
797
+ },
798
+ {
799
+ "epoch": 0.7179818887451488,
800
+ "grad_norm": 0.2839801013469696,
801
+ "learning_rate": 0.00013951442646023928,
802
+ "loss": 0.4332,
803
+ "step": 2220
804
+ },
805
+ {
806
+ "epoch": 0.7244501940491591,
807
+ "grad_norm": 0.27037495374679565,
808
+ "learning_rate": 0.00013881069669247008,
809
+ "loss": 0.4382,
810
+ "step": 2240
811
+ },
812
+ {
813
+ "epoch": 0.7309184993531694,
814
+ "grad_norm": 0.34917208552360535,
815
+ "learning_rate": 0.00013810696692470093,
816
+ "loss": 0.4251,
817
+ "step": 2260
818
+ },
819
+ {
820
+ "epoch": 0.7373868046571799,
821
+ "grad_norm": 0.39801281690597534,
822
+ "learning_rate": 0.00013740323715693176,
823
+ "loss": 0.4432,
824
+ "step": 2280
825
+ },
826
+ {
827
+ "epoch": 0.7438551099611902,
828
+ "grad_norm": 0.8932256102561951,
829
+ "learning_rate": 0.00013669950738916259,
830
+ "loss": 0.4577,
831
+ "step": 2300
832
+ },
833
+ {
834
+ "epoch": 0.7503234152652005,
835
+ "grad_norm": 0.35125109553337097,
836
+ "learning_rate": 0.00013599577762139339,
837
+ "loss": 0.4502,
838
+ "step": 2320
839
+ },
840
+ {
841
+ "epoch": 0.7567917205692108,
842
+ "grad_norm": 0.3012046813964844,
843
+ "learning_rate": 0.0001352920478536242,
844
+ "loss": 0.4513,
845
+ "step": 2340
846
+ },
847
+ {
848
+ "epoch": 0.7632600258732212,
849
+ "grad_norm": 0.3242063820362091,
850
+ "learning_rate": 0.00013458831808585504,
851
+ "loss": 0.4559,
852
+ "step": 2360
853
+ },
854
+ {
855
+ "epoch": 0.7697283311772316,
856
+ "grad_norm": 0.3523860275745392,
857
+ "learning_rate": 0.00013388458831808587,
858
+ "loss": 0.4434,
859
+ "step": 2380
860
+ },
861
+ {
862
+ "epoch": 0.7761966364812419,
863
+ "grad_norm": 0.29160013794898987,
864
+ "learning_rate": 0.0001331808585503167,
865
+ "loss": 0.44,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 0.7826649417852523,
870
+ "grad_norm": 0.3047971725463867,
871
+ "learning_rate": 0.0001324771287825475,
872
+ "loss": 0.4336,
873
+ "step": 2420
874
+ },
875
+ {
876
+ "epoch": 0.7891332470892626,
877
+ "grad_norm": 0.3192126750946045,
878
+ "learning_rate": 0.00013177339901477832,
879
+ "loss": 0.4495,
880
+ "step": 2440
881
+ },
882
+ {
883
+ "epoch": 0.795601552393273,
884
+ "grad_norm": 0.37997299432754517,
885
+ "learning_rate": 0.00013106966924700917,
886
+ "loss": 0.4685,
887
+ "step": 2460
888
+ },
889
+ {
890
+ "epoch": 0.8020698576972833,
891
+ "grad_norm": 0.2811839282512665,
892
+ "learning_rate": 0.00013036593947924,
893
+ "loss": 0.4206,
894
+ "step": 2480
895
+ },
896
+ {
897
+ "epoch": 0.8085381630012937,
898
+ "grad_norm": 0.3121369779109955,
899
+ "learning_rate": 0.0001296622097114708,
900
+ "loss": 0.425,
901
+ "step": 2500
902
+ },
903
+ {
904
+ "epoch": 0.815006468305304,
905
+ "grad_norm": 0.3134746253490448,
906
+ "learning_rate": 0.00012895847994370162,
907
+ "loss": 0.4498,
908
+ "step": 2520
909
+ },
910
+ {
911
+ "epoch": 0.8214747736093143,
912
+ "grad_norm": 0.36752375960350037,
913
+ "learning_rate": 0.00012825475017593245,
914
+ "loss": 0.4288,
915
+ "step": 2540
916
+ },
917
+ {
918
+ "epoch": 0.8279430789133247,
919
+ "grad_norm": 0.3915141224861145,
920
+ "learning_rate": 0.00012755102040816328,
921
+ "loss": 0.4605,
922
+ "step": 2560
923
+ },
924
+ {
925
+ "epoch": 0.8344113842173351,
926
+ "grad_norm": 0.3662431538105011,
927
+ "learning_rate": 0.00012684729064039408,
928
+ "loss": 0.4194,
929
+ "step": 2580
930
+ },
931
+ {
932
+ "epoch": 0.8408796895213454,
933
+ "grad_norm": 0.3387574851512909,
934
+ "learning_rate": 0.0001261435608726249,
935
+ "loss": 0.4258,
936
+ "step": 2600
937
+ },
938
+ {
939
+ "epoch": 0.8473479948253557,
940
+ "grad_norm": 0.38025060296058655,
941
+ "learning_rate": 0.00012543983110485573,
942
+ "loss": 0.4291,
943
+ "step": 2620
944
+ },
945
+ {
946
+ "epoch": 0.8538163001293662,
947
+ "grad_norm": 0.3517441749572754,
948
+ "learning_rate": 0.00012473610133708658,
949
+ "loss": 0.4384,
950
+ "step": 2640
951
+ },
952
+ {
953
+ "epoch": 0.8602846054333765,
954
+ "grad_norm": 0.33037152886390686,
955
+ "learning_rate": 0.0001240323715693174,
956
+ "loss": 0.4452,
957
+ "step": 2660
958
+ },
959
+ {
960
+ "epoch": 0.8667529107373868,
961
+ "grad_norm": 0.3645654618740082,
962
+ "learning_rate": 0.0001233286418015482,
963
+ "loss": 0.3992,
964
+ "step": 2680
965
+ },
966
+ {
967
+ "epoch": 0.8732212160413971,
968
+ "grad_norm": 0.32396814227104187,
969
+ "learning_rate": 0.00012262491203377904,
970
+ "loss": 0.4,
971
+ "step": 2700
972
+ },
973
+ {
974
+ "epoch": 0.8796895213454075,
975
+ "grad_norm": 0.3377726078033447,
976
+ "learning_rate": 0.00012192118226600986,
977
+ "loss": 0.4306,
978
+ "step": 2720
979
+ },
980
+ {
981
+ "epoch": 0.8861578266494179,
982
+ "grad_norm": 0.32345983386039734,
983
+ "learning_rate": 0.00012121745249824067,
984
+ "loss": 0.4225,
985
+ "step": 2740
986
+ },
987
+ {
988
+ "epoch": 0.8926261319534282,
989
+ "grad_norm": 0.34778884053230286,
990
+ "learning_rate": 0.0001205137227304715,
991
+ "loss": 0.4256,
992
+ "step": 2760
993
+ },
994
+ {
995
+ "epoch": 0.8990944372574385,
996
+ "grad_norm": 0.3861066997051239,
997
+ "learning_rate": 0.00011980999296270233,
998
+ "loss": 0.4465,
999
+ "step": 2780
1000
+ },
1001
+ {
1002
+ "epoch": 0.9055627425614489,
1003
+ "grad_norm": 0.3882784843444824,
1004
+ "learning_rate": 0.00011910626319493314,
1005
+ "loss": 0.4026,
1006
+ "step": 2800
1007
+ },
1008
+ {
1009
+ "epoch": 0.9120310478654593,
1010
+ "grad_norm": 0.4380407929420471,
1011
+ "learning_rate": 0.00011840253342716397,
1012
+ "loss": 0.4375,
1013
+ "step": 2820
1014
+ },
1015
+ {
1016
+ "epoch": 0.9184993531694696,
1017
+ "grad_norm": 0.35943394899368286,
1018
+ "learning_rate": 0.00011769880365939481,
1019
+ "loss": 0.4337,
1020
+ "step": 2840
1021
+ },
1022
+ {
1023
+ "epoch": 0.92496765847348,
1024
+ "grad_norm": 0.3862978219985962,
1025
+ "learning_rate": 0.00011699507389162563,
1026
+ "loss": 0.3949,
1027
+ "step": 2860
1028
+ },
1029
+ {
1030
+ "epoch": 0.9314359637774903,
1031
+ "grad_norm": 0.34592369198799133,
1032
+ "learning_rate": 0.00011629134412385645,
1033
+ "loss": 0.3798,
1034
+ "step": 2880
1035
+ },
1036
+ {
1037
+ "epoch": 0.9379042690815006,
1038
+ "grad_norm": 0.34942877292633057,
1039
+ "learning_rate": 0.00011558761435608727,
1040
+ "loss": 0.4235,
1041
+ "step": 2900
1042
+ },
1043
+ {
1044
+ "epoch": 0.944372574385511,
1045
+ "grad_norm": 0.28018903732299805,
1046
+ "learning_rate": 0.00011488388458831809,
1047
+ "loss": 0.4353,
1048
+ "step": 2920
1049
+ },
1050
+ {
1051
+ "epoch": 0.9508408796895214,
1052
+ "grad_norm": 0.3505672812461853,
1053
+ "learning_rate": 0.00011418015482054891,
1054
+ "loss": 0.3954,
1055
+ "step": 2940
1056
+ },
1057
+ {
1058
+ "epoch": 0.9573091849935317,
1059
+ "grad_norm": 0.3264220654964447,
1060
+ "learning_rate": 0.00011347642505277974,
1061
+ "loss": 0.4422,
1062
+ "step": 2960
1063
+ },
1064
+ {
1065
+ "epoch": 0.963777490297542,
1066
+ "grad_norm": 0.36382535099983215,
1067
+ "learning_rate": 0.00011277269528501055,
1068
+ "loss": 0.4116,
1069
+ "step": 2980
1070
+ },
1071
+ {
1072
+ "epoch": 0.9702457956015524,
1073
+ "grad_norm": 0.3051494359970093,
1074
+ "learning_rate": 0.00011206896551724138,
1075
+ "loss": 0.4135,
1076
+ "step": 3000
1077
+ },
1078
+ {
1079
+ "epoch": 0.9702457956015524,
1080
+ "eval_loss": 0.5850913524627686,
1081
+ "eval_runtime": 967.6256,
1082
+ "eval_samples_per_second": 6.332,
1083
+ "eval_steps_per_second": 0.792,
1084
+ "step": 3000
1085
+ },
1086
+ {
1087
+ "epoch": 0.9767141009055628,
1088
+ "grad_norm": 0.34533271193504333,
1089
+ "learning_rate": 0.00011136523574947222,
1090
+ "loss": 0.4312,
1091
+ "step": 3020
1092
+ },
1093
+ {
1094
+ "epoch": 0.9831824062095731,
1095
+ "grad_norm": 0.35221609473228455,
1096
+ "learning_rate": 0.00011066150598170305,
1097
+ "loss": 0.4245,
1098
+ "step": 3040
1099
+ },
1100
+ {
1101
+ "epoch": 0.9896507115135834,
1102
+ "grad_norm": 0.3893713057041168,
1103
+ "learning_rate": 0.00010995777621393386,
1104
+ "loss": 0.4048,
1105
+ "step": 3060
1106
+ },
1107
+ {
1108
+ "epoch": 0.9961190168175937,
1109
+ "grad_norm": 0.3586278557777405,
1110
+ "learning_rate": 0.00010925404644616468,
1111
+ "loss": 0.4159,
1112
+ "step": 3080
1113
+ },
1114
+ {
1115
+ "epoch": 1.0025873221216042,
1116
+ "grad_norm": 0.32620444893836975,
1117
+ "learning_rate": 0.0001085503166783955,
1118
+ "loss": 0.3741,
1119
+ "step": 3100
1120
+ },
1121
+ {
1122
+ "epoch": 1.0090556274256144,
1123
+ "grad_norm": 0.33812645077705383,
1124
+ "learning_rate": 0.00010784658691062632,
1125
+ "loss": 0.3694,
1126
+ "step": 3120
1127
+ },
1128
+ {
1129
+ "epoch": 1.0155239327296248,
1130
+ "grad_norm": 0.39038708806037903,
1131
+ "learning_rate": 0.00010714285714285715,
1132
+ "loss": 0.3882,
1133
+ "step": 3140
1134
+ },
1135
+ {
1136
+ "epoch": 1.0219922380336353,
1137
+ "grad_norm": 0.3209969997406006,
1138
+ "learning_rate": 0.00010643912737508796,
1139
+ "loss": 0.3437,
1140
+ "step": 3160
1141
+ },
1142
+ {
1143
+ "epoch": 1.0284605433376455,
1144
+ "grad_norm": 0.3705078661441803,
1145
+ "learning_rate": 0.00010573539760731879,
1146
+ "loss": 0.3568,
1147
+ "step": 3180
1148
+ },
1149
+ {
1150
+ "epoch": 1.034928848641656,
1151
+ "grad_norm": 0.3562203645706177,
1152
+ "learning_rate": 0.0001050316678395496,
1153
+ "loss": 0.3916,
1154
+ "step": 3200
1155
+ },
1156
+ {
1157
+ "epoch": 1.0413971539456663,
1158
+ "grad_norm": 0.3724074363708496,
1159
+ "learning_rate": 0.00010432793807178046,
1160
+ "loss": 0.3866,
1161
+ "step": 3220
1162
+ },
1163
+ {
1164
+ "epoch": 1.0478654592496766,
1165
+ "grad_norm": 0.40453559160232544,
1166
+ "learning_rate": 0.00010362420830401127,
1167
+ "loss": 0.368,
1168
+ "step": 3240
1169
+ },
1170
+ {
1171
+ "epoch": 1.054333764553687,
1172
+ "grad_norm": 0.3316144049167633,
1173
+ "learning_rate": 0.0001029204785362421,
1174
+ "loss": 0.3696,
1175
+ "step": 3260
1176
+ },
1177
+ {
1178
+ "epoch": 1.0608020698576972,
1179
+ "grad_norm": 0.3503650724887848,
1180
+ "learning_rate": 0.00010221674876847291,
1181
+ "loss": 0.3644,
1182
+ "step": 3280
1183
+ },
1184
+ {
1185
+ "epoch": 1.0672703751617076,
1186
+ "grad_norm": 0.3136923909187317,
1187
+ "learning_rate": 0.00010151301900070374,
1188
+ "loss": 0.3414,
1189
+ "step": 3300
1190
+ },
1191
+ {
1192
+ "epoch": 1.073738680465718,
1193
+ "grad_norm": 0.3573864996433258,
1194
+ "learning_rate": 0.00010080928923293456,
1195
+ "loss": 0.3586,
1196
+ "step": 3320
1197
+ },
1198
+ {
1199
+ "epoch": 1.0802069857697283,
1200
+ "grad_norm": 0.26215964555740356,
1201
+ "learning_rate": 0.00010010555946516538,
1202
+ "loss": 0.356,
1203
+ "step": 3340
1204
+ },
1205
+ {
1206
+ "epoch": 1.0866752910737387,
1207
+ "grad_norm": 0.3103466331958771,
1208
+ "learning_rate": 9.940182969739622e-05,
1209
+ "loss": 0.3478,
1210
+ "step": 3360
1211
+ },
1212
+ {
1213
+ "epoch": 1.093143596377749,
1214
+ "grad_norm": 0.3601842224597931,
1215
+ "learning_rate": 9.869809992962703e-05,
1216
+ "loss": 0.3635,
1217
+ "step": 3380
1218
+ },
1219
+ {
1220
+ "epoch": 1.0996119016817594,
1221
+ "grad_norm": 0.3715918958187103,
1222
+ "learning_rate": 9.799437016185785e-05,
1223
+ "loss": 0.3489,
1224
+ "step": 3400
1225
+ },
1226
+ {
1227
+ "epoch": 1.1060802069857698,
1228
+ "grad_norm": 0.4216213524341583,
1229
+ "learning_rate": 9.729064039408867e-05,
1230
+ "loss": 0.3461,
1231
+ "step": 3420
1232
+ },
1233
+ {
1234
+ "epoch": 1.11254851228978,
1235
+ "grad_norm": 0.3878704905509949,
1236
+ "learning_rate": 9.65869106263195e-05,
1237
+ "loss": 0.3548,
1238
+ "step": 3440
1239
+ },
1240
+ {
1241
+ "epoch": 1.1190168175937905,
1242
+ "grad_norm": 0.36658766865730286,
1243
+ "learning_rate": 9.588318085855032e-05,
1244
+ "loss": 0.3395,
1245
+ "step": 3460
1246
+ },
1247
+ {
1248
+ "epoch": 1.1254851228978007,
1249
+ "grad_norm": 0.38134855031967163,
1250
+ "learning_rate": 9.517945109078115e-05,
1251
+ "loss": 0.3829,
1252
+ "step": 3480
1253
+ },
1254
+ {
1255
+ "epoch": 1.131953428201811,
1256
+ "grad_norm": 0.42683905363082886,
1257
+ "learning_rate": 9.447572132301197e-05,
1258
+ "loss": 0.3914,
1259
+ "step": 3500
1260
+ },
1261
+ {
1262
+ "epoch": 1.1384217335058215,
1263
+ "grad_norm": 0.3340086042881012,
1264
+ "learning_rate": 9.377199155524279e-05,
1265
+ "loss": 0.3301,
1266
+ "step": 3520
1267
+ },
1268
+ {
1269
+ "epoch": 1.1448900388098318,
1270
+ "grad_norm": 0.37897104024887085,
1271
+ "learning_rate": 9.306826178747361e-05,
1272
+ "loss": 0.3536,
1273
+ "step": 3540
1274
+ },
1275
+ {
1276
+ "epoch": 1.1513583441138422,
1277
+ "grad_norm": 0.361061692237854,
1278
+ "learning_rate": 9.236453201970444e-05,
1279
+ "loss": 0.338,
1280
+ "step": 3560
1281
+ },
1282
+ {
1283
+ "epoch": 1.1578266494178524,
1284
+ "grad_norm": 0.34730997681617737,
1285
+ "learning_rate": 9.166080225193527e-05,
1286
+ "loss": 0.3679,
1287
+ "step": 3580
1288
+ },
1289
+ {
1290
+ "epoch": 1.1642949547218628,
1291
+ "grad_norm": 0.29455122351646423,
1292
+ "learning_rate": 9.095707248416608e-05,
1293
+ "loss": 0.3602,
1294
+ "step": 3600
1295
+ },
1296
+ {
1297
+ "epoch": 1.1707632600258733,
1298
+ "grad_norm": 0.33159056305885315,
1299
+ "learning_rate": 9.02533427163969e-05,
1300
+ "loss": 0.3359,
1301
+ "step": 3620
1302
+ },
1303
+ {
1304
+ "epoch": 1.1772315653298835,
1305
+ "grad_norm": 0.48726096749305725,
1306
+ "learning_rate": 8.954961294862773e-05,
1307
+ "loss": 0.3879,
1308
+ "step": 3640
1309
+ },
1310
+ {
1311
+ "epoch": 1.183699870633894,
1312
+ "grad_norm": 0.3182481825351715,
1313
+ "learning_rate": 8.884588318085856e-05,
1314
+ "loss": 0.3392,
1315
+ "step": 3660
1316
+ },
1317
+ {
1318
+ "epoch": 1.1901681759379044,
1319
+ "grad_norm": 0.3488766551017761,
1320
+ "learning_rate": 8.814215341308939e-05,
1321
+ "loss": 0.3419,
1322
+ "step": 3680
1323
+ },
1324
+ {
1325
+ "epoch": 1.1966364812419146,
1326
+ "grad_norm": 0.34345540404319763,
1327
+ "learning_rate": 8.74384236453202e-05,
1328
+ "loss": 0.3598,
1329
+ "step": 3700
1330
+ },
1331
+ {
1332
+ "epoch": 1.203104786545925,
1333
+ "grad_norm": 0.37642374634742737,
1334
+ "learning_rate": 8.673469387755102e-05,
1335
+ "loss": 0.3578,
1336
+ "step": 3720
1337
+ },
1338
+ {
1339
+ "epoch": 1.2095730918499352,
1340
+ "grad_norm": 0.3726005554199219,
1341
+ "learning_rate": 8.603096410978185e-05,
1342
+ "loss": 0.3547,
1343
+ "step": 3740
1344
+ },
1345
+ {
1346
+ "epoch": 1.2160413971539457,
1347
+ "grad_norm": 0.36603832244873047,
1348
+ "learning_rate": 8.532723434201268e-05,
1349
+ "loss": 0.3831,
1350
+ "step": 3760
1351
+ },
1352
+ {
1353
+ "epoch": 1.222509702457956,
1354
+ "grad_norm": 0.345103919506073,
1355
+ "learning_rate": 8.462350457424349e-05,
1356
+ "loss": 0.3563,
1357
+ "step": 3780
1358
+ },
1359
+ {
1360
+ "epoch": 1.2289780077619663,
1361
+ "grad_norm": 0.31071504950523376,
1362
+ "learning_rate": 8.391977480647432e-05,
1363
+ "loss": 0.3263,
1364
+ "step": 3800
1365
+ },
1366
+ {
1367
+ "epoch": 1.2354463130659767,
1368
+ "grad_norm": 0.3534263074398041,
1369
+ "learning_rate": 8.321604503870513e-05,
1370
+ "loss": 0.3591,
1371
+ "step": 3820
1372
+ },
1373
+ {
1374
+ "epoch": 1.2419146183699872,
1375
+ "grad_norm": 0.3568720817565918,
1376
+ "learning_rate": 8.251231527093597e-05,
1377
+ "loss": 0.3815,
1378
+ "step": 3840
1379
+ },
1380
+ {
1381
+ "epoch": 1.2483829236739974,
1382
+ "grad_norm": 0.3450963795185089,
1383
+ "learning_rate": 8.18085855031668e-05,
1384
+ "loss": 0.3481,
1385
+ "step": 3860
1386
+ },
1387
+ {
1388
+ "epoch": 1.2548512289780078,
1389
+ "grad_norm": 0.3554648160934448,
1390
+ "learning_rate": 8.110485573539761e-05,
1391
+ "loss": 0.3364,
1392
+ "step": 3880
1393
+ },
1394
+ {
1395
+ "epoch": 1.261319534282018,
1396
+ "grad_norm": 0.4188801348209381,
1397
+ "learning_rate": 8.040112596762844e-05,
1398
+ "loss": 0.3686,
1399
+ "step": 3900
1400
+ },
1401
+ {
1402
+ "epoch": 1.2677878395860285,
1403
+ "grad_norm": 0.3129253089427948,
1404
+ "learning_rate": 7.969739619985925e-05,
1405
+ "loss": 0.3664,
1406
+ "step": 3920
1407
+ },
1408
+ {
1409
+ "epoch": 1.274256144890039,
1410
+ "grad_norm": 0.3885388970375061,
1411
+ "learning_rate": 7.899366643209009e-05,
1412
+ "loss": 0.3415,
1413
+ "step": 3940
1414
+ },
1415
+ {
1416
+ "epoch": 1.2807244501940491,
1417
+ "grad_norm": 0.4131453335285187,
1418
+ "learning_rate": 7.82899366643209e-05,
1419
+ "loss": 0.3552,
1420
+ "step": 3960
1421
+ },
1422
+ {
1423
+ "epoch": 1.2871927554980596,
1424
+ "grad_norm": 0.39486396312713623,
1425
+ "learning_rate": 7.758620689655173e-05,
1426
+ "loss": 0.366,
1427
+ "step": 3980
1428
+ },
1429
+ {
1430
+ "epoch": 1.2936610608020698,
1431
+ "grad_norm": 0.3580794930458069,
1432
+ "learning_rate": 7.688247712878254e-05,
1433
+ "loss": 0.3585,
1434
+ "step": 4000
1435
+ },
1436
+ {
1437
+ "epoch": 1.2936610608020698,
1438
+ "eval_loss": 0.5952740907669067,
1439
+ "eval_runtime": 967.8131,
1440
+ "eval_samples_per_second": 6.331,
1441
+ "eval_steps_per_second": 0.791,
1442
+ "step": 4000
1443
+ },
1444
+ {
1445
+ "epoch": 1.3001293661060802,
1446
+ "grad_norm": 0.3903474509716034,
1447
+ "learning_rate": 7.617874736101337e-05,
1448
+ "loss": 0.3347,
1449
+ "step": 4020
1450
+ },
1451
+ {
1452
+ "epoch": 1.3065976714100906,
1453
+ "grad_norm": 0.4154898226261139,
1454
+ "learning_rate": 7.54750175932442e-05,
1455
+ "loss": 0.3314,
1456
+ "step": 4040
1457
+ },
1458
+ {
1459
+ "epoch": 1.3130659767141009,
1460
+ "grad_norm": 0.33763521909713745,
1461
+ "learning_rate": 7.477128782547502e-05,
1462
+ "loss": 0.3552,
1463
+ "step": 4060
1464
+ },
1465
+ {
1466
+ "epoch": 1.3195342820181113,
1467
+ "grad_norm": 0.5618635416030884,
1468
+ "learning_rate": 7.406755805770585e-05,
1469
+ "loss": 0.363,
1470
+ "step": 4080
1471
+ },
1472
+ {
1473
+ "epoch": 1.3260025873221215,
1474
+ "grad_norm": 0.37114137411117554,
1475
+ "learning_rate": 7.336382828993666e-05,
1476
+ "loss": 0.3421,
1477
+ "step": 4100
1478
+ },
1479
+ {
1480
+ "epoch": 1.332470892626132,
1481
+ "grad_norm": 0.34323641657829285,
1482
+ "learning_rate": 7.266009852216749e-05,
1483
+ "loss": 0.3659,
1484
+ "step": 4120
1485
+ },
1486
+ {
1487
+ "epoch": 1.3389391979301424,
1488
+ "grad_norm": 0.3527199625968933,
1489
+ "learning_rate": 7.195636875439831e-05,
1490
+ "loss": 0.3412,
1491
+ "step": 4140
1492
+ },
1493
+ {
1494
+ "epoch": 1.3454075032341526,
1495
+ "grad_norm": 0.3413406014442444,
1496
+ "learning_rate": 7.125263898662914e-05,
1497
+ "loss": 0.338,
1498
+ "step": 4160
1499
+ },
1500
+ {
1501
+ "epoch": 1.351875808538163,
1502
+ "grad_norm": 0.34787651896476746,
1503
+ "learning_rate": 7.054890921885995e-05,
1504
+ "loss": 0.3681,
1505
+ "step": 4180
1506
+ },
1507
+ {
1508
+ "epoch": 1.3583441138421732,
1509
+ "grad_norm": 0.2709845006465912,
1510
+ "learning_rate": 6.984517945109078e-05,
1511
+ "loss": 0.325,
1512
+ "step": 4200
1513
+ },
1514
+ {
1515
+ "epoch": 1.3648124191461837,
1516
+ "grad_norm": 0.30966514348983765,
1517
+ "learning_rate": 6.91414496833216e-05,
1518
+ "loss": 0.3677,
1519
+ "step": 4220
1520
+ },
1521
+ {
1522
+ "epoch": 1.371280724450194,
1523
+ "grad_norm": 0.35980409383773804,
1524
+ "learning_rate": 6.843771991555243e-05,
1525
+ "loss": 0.335,
1526
+ "step": 4240
1527
+ },
1528
+ {
1529
+ "epoch": 1.3777490297542043,
1530
+ "grad_norm": 0.5173650979995728,
1531
+ "learning_rate": 6.773399014778326e-05,
1532
+ "loss": 0.348,
1533
+ "step": 4260
1534
+ },
1535
+ {
1536
+ "epoch": 1.3842173350582148,
1537
+ "grad_norm": 0.3048713207244873,
1538
+ "learning_rate": 6.703026038001407e-05,
1539
+ "loss": 0.3323,
1540
+ "step": 4280
1541
+ },
1542
+ {
1543
+ "epoch": 1.390685640362225,
1544
+ "grad_norm": 0.3427985906600952,
1545
+ "learning_rate": 6.63265306122449e-05,
1546
+ "loss": 0.3571,
1547
+ "step": 4300
1548
+ },
1549
+ {
1550
+ "epoch": 1.3971539456662354,
1551
+ "grad_norm": 0.406012624502182,
1552
+ "learning_rate": 6.562280084447573e-05,
1553
+ "loss": 0.3546,
1554
+ "step": 4320
1555
+ },
1556
+ {
1557
+ "epoch": 1.4036222509702458,
1558
+ "grad_norm": 0.31981325149536133,
1559
+ "learning_rate": 6.491907107670655e-05,
1560
+ "loss": 0.3281,
1561
+ "step": 4340
1562
+ },
1563
+ {
1564
+ "epoch": 1.4100905562742563,
1565
+ "grad_norm": 0.3936704099178314,
1566
+ "learning_rate": 6.421534130893737e-05,
1567
+ "loss": 0.3444,
1568
+ "step": 4360
1569
+ },
1570
+ {
1571
+ "epoch": 1.4165588615782665,
1572
+ "grad_norm": 0.3658936023712158,
1573
+ "learning_rate": 6.351161154116819e-05,
1574
+ "loss": 0.339,
1575
+ "step": 4380
1576
+ },
1577
+ {
1578
+ "epoch": 1.4230271668822767,
1579
+ "grad_norm": 0.32523271441459656,
1580
+ "learning_rate": 6.280788177339902e-05,
1581
+ "loss": 0.3311,
1582
+ "step": 4400
1583
+ },
1584
+ {
1585
+ "epoch": 1.4294954721862871,
1586
+ "grad_norm": 0.4120417833328247,
1587
+ "learning_rate": 6.210415200562984e-05,
1588
+ "loss": 0.3342,
1589
+ "step": 4420
1590
+ },
1591
+ {
1592
+ "epoch": 1.4359637774902976,
1593
+ "grad_norm": 0.35293978452682495,
1594
+ "learning_rate": 6.140042223786067e-05,
1595
+ "loss": 0.3561,
1596
+ "step": 4440
1597
+ },
1598
+ {
1599
+ "epoch": 1.442432082794308,
1600
+ "grad_norm": 0.33231040835380554,
1601
+ "learning_rate": 6.0696692470091484e-05,
1602
+ "loss": 0.3573,
1603
+ "step": 4460
1604
+ },
1605
+ {
1606
+ "epoch": 1.4489003880983182,
1607
+ "grad_norm": 0.4345926344394684,
1608
+ "learning_rate": 5.999296270232231e-05,
1609
+ "loss": 0.3364,
1610
+ "step": 4480
1611
+ },
1612
+ {
1613
+ "epoch": 1.4553686934023287,
1614
+ "grad_norm": 0.34958288073539734,
1615
+ "learning_rate": 5.928923293455313e-05,
1616
+ "loss": 0.3162,
1617
+ "step": 4500
1618
+ },
1619
+ {
1620
+ "epoch": 1.4618369987063389,
1621
+ "grad_norm": 0.4669477045536041,
1622
+ "learning_rate": 5.8585503166783964e-05,
1623
+ "loss": 0.3455,
1624
+ "step": 4520
1625
+ },
1626
+ {
1627
+ "epoch": 1.4683053040103493,
1628
+ "grad_norm": 0.32660993933677673,
1629
+ "learning_rate": 5.788177339901478e-05,
1630
+ "loss": 0.3228,
1631
+ "step": 4540
1632
+ },
1633
+ {
1634
+ "epoch": 1.4747736093143597,
1635
+ "grad_norm": 0.3654051423072815,
1636
+ "learning_rate": 5.71780436312456e-05,
1637
+ "loss": 0.3293,
1638
+ "step": 4560
1639
+ },
1640
+ {
1641
+ "epoch": 1.48124191461837,
1642
+ "grad_norm": 0.3496916592121124,
1643
+ "learning_rate": 5.647431386347642e-05,
1644
+ "loss": 0.3382,
1645
+ "step": 4580
1646
+ },
1647
+ {
1648
+ "epoch": 1.4877102199223804,
1649
+ "grad_norm": 0.3890388309955597,
1650
+ "learning_rate": 5.577058409570724e-05,
1651
+ "loss": 0.3468,
1652
+ "step": 4600
1653
+ },
1654
+ {
1655
+ "epoch": 1.4941785252263906,
1656
+ "grad_norm": 0.3924092650413513,
1657
+ "learning_rate": 5.5066854327938076e-05,
1658
+ "loss": 0.3467,
1659
+ "step": 4620
1660
+ },
1661
+ {
1662
+ "epoch": 1.500646830530401,
1663
+ "grad_norm": 0.4015033543109894,
1664
+ "learning_rate": 5.4363124560168896e-05,
1665
+ "loss": 0.3478,
1666
+ "step": 4640
1667
+ },
1668
+ {
1669
+ "epoch": 1.5071151358344115,
1670
+ "grad_norm": 0.34742093086242676,
1671
+ "learning_rate": 5.365939479239972e-05,
1672
+ "loss": 0.3252,
1673
+ "step": 4660
1674
+ },
1675
+ {
1676
+ "epoch": 1.5135834411384217,
1677
+ "grad_norm": 0.3477989435195923,
1678
+ "learning_rate": 5.295566502463054e-05,
1679
+ "loss": 0.3627,
1680
+ "step": 4680
1681
+ },
1682
+ {
1683
+ "epoch": 1.5200517464424321,
1684
+ "grad_norm": 0.4194832146167755,
1685
+ "learning_rate": 5.2251935256861375e-05,
1686
+ "loss": 0.3403,
1687
+ "step": 4700
1688
+ },
1689
+ {
1690
+ "epoch": 1.5265200517464423,
1691
+ "grad_norm": 0.37723594903945923,
1692
+ "learning_rate": 5.1548205489092195e-05,
1693
+ "loss": 0.3424,
1694
+ "step": 4720
1695
+ },
1696
+ {
1697
+ "epoch": 1.5329883570504528,
1698
+ "grad_norm": 0.3695043921470642,
1699
+ "learning_rate": 5.0844475721323015e-05,
1700
+ "loss": 0.3339,
1701
+ "step": 4740
1702
+ },
1703
+ {
1704
+ "epoch": 1.5394566623544632,
1705
+ "grad_norm": 0.35015031695365906,
1706
+ "learning_rate": 5.0140745953553834e-05,
1707
+ "loss": 0.3262,
1708
+ "step": 4760
1709
+ },
1710
+ {
1711
+ "epoch": 1.5459249676584734,
1712
+ "grad_norm": 0.39946597814559937,
1713
+ "learning_rate": 4.943701618578466e-05,
1714
+ "loss": 0.3649,
1715
+ "step": 4780
1716
+ },
1717
+ {
1718
+ "epoch": 1.5523932729624839,
1719
+ "grad_norm": 0.3246670961380005,
1720
+ "learning_rate": 4.873328641801548e-05,
1721
+ "loss": 0.3241,
1722
+ "step": 4800
1723
+ },
1724
+ {
1725
+ "epoch": 1.558861578266494,
1726
+ "grad_norm": 0.41913464665412903,
1727
+ "learning_rate": 4.802955665024631e-05,
1728
+ "loss": 0.3486,
1729
+ "step": 4820
1730
+ },
1731
+ {
1732
+ "epoch": 1.5653298835705045,
1733
+ "grad_norm": 0.38162580132484436,
1734
+ "learning_rate": 4.7325826882477134e-05,
1735
+ "loss": 0.3571,
1736
+ "step": 4840
1737
+ },
1738
+ {
1739
+ "epoch": 1.571798188874515,
1740
+ "grad_norm": 0.34426695108413696,
1741
+ "learning_rate": 4.6622097114707953e-05,
1742
+ "loss": 0.3232,
1743
+ "step": 4860
1744
+ },
1745
+ {
1746
+ "epoch": 1.5782664941785254,
1747
+ "grad_norm": 0.3782813847064972,
1748
+ "learning_rate": 4.591836734693878e-05,
1749
+ "loss": 0.346,
1750
+ "step": 4880
1751
+ },
1752
+ {
1753
+ "epoch": 1.5847347994825356,
1754
+ "grad_norm": 0.42736709117889404,
1755
+ "learning_rate": 4.52146375791696e-05,
1756
+ "loss": 0.3275,
1757
+ "step": 4900
1758
+ },
1759
+ {
1760
+ "epoch": 1.5912031047865458,
1761
+ "grad_norm": 0.37077978253364563,
1762
+ "learning_rate": 4.4510907811400426e-05,
1763
+ "loss": 0.3312,
1764
+ "step": 4920
1765
+ },
1766
+ {
1767
+ "epoch": 1.5976714100905562,
1768
+ "grad_norm": 0.3196958303451538,
1769
+ "learning_rate": 4.3807178043631246e-05,
1770
+ "loss": 0.3156,
1771
+ "step": 4940
1772
+ },
1773
+ {
1774
+ "epoch": 1.6041397153945667,
1775
+ "grad_norm": 0.36680254340171814,
1776
+ "learning_rate": 4.3103448275862066e-05,
1777
+ "loss": 0.3293,
1778
+ "step": 4960
1779
+ },
1780
+ {
1781
+ "epoch": 1.6106080206985771,
1782
+ "grad_norm": 0.44455257058143616,
1783
+ "learning_rate": 4.239971850809289e-05,
1784
+ "loss": 0.3512,
1785
+ "step": 4980
1786
+ },
1787
+ {
1788
+ "epoch": 1.6170763260025873,
1789
+ "grad_norm": 0.37418290972709656,
1790
+ "learning_rate": 4.169598874032372e-05,
1791
+ "loss": 0.3422,
1792
+ "step": 5000
1793
+ },
1794
+ {
1795
+ "epoch": 1.6170763260025873,
1796
+ "eval_loss": 0.5881822109222412,
1797
+ "eval_runtime": 967.8912,
1798
+ "eval_samples_per_second": 6.33,
1799
+ "eval_steps_per_second": 0.791,
1800
+ "step": 5000
1801
+ },
1802
+ {
1803
+ "epoch": 1.6235446313065975,
1804
+ "grad_norm": 0.3209955096244812,
1805
+ "learning_rate": 4.0992258972554545e-05,
1806
+ "loss": 0.3158,
1807
+ "step": 5020
1808
+ },
1809
+ {
1810
+ "epoch": 1.630012936610608,
1811
+ "grad_norm": 0.31715771555900574,
1812
+ "learning_rate": 4.0288529204785365e-05,
1813
+ "loss": 0.3167,
1814
+ "step": 5040
1815
+ },
1816
+ {
1817
+ "epoch": 1.6364812419146184,
1818
+ "grad_norm": 0.3968450725078583,
1819
+ "learning_rate": 3.958479943701619e-05,
1820
+ "loss": 0.3219,
1821
+ "step": 5060
1822
+ },
1823
+ {
1824
+ "epoch": 1.6429495472186288,
1825
+ "grad_norm": 0.3187231123447418,
1826
+ "learning_rate": 3.888106966924701e-05,
1827
+ "loss": 0.3288,
1828
+ "step": 5080
1829
+ },
1830
+ {
1831
+ "epoch": 1.649417852522639,
1832
+ "grad_norm": 0.35491153597831726,
1833
+ "learning_rate": 3.817733990147783e-05,
1834
+ "loss": 0.3298,
1835
+ "step": 5100
1836
+ },
1837
+ {
1838
+ "epoch": 1.6558861578266493,
1839
+ "grad_norm": 0.4044147729873657,
1840
+ "learning_rate": 3.747361013370866e-05,
1841
+ "loss": 0.3505,
1842
+ "step": 5120
1843
+ },
1844
+ {
1845
+ "epoch": 1.6623544631306597,
1846
+ "grad_norm": 0.33881333470344543,
1847
+ "learning_rate": 3.676988036593948e-05,
1848
+ "loss": 0.3269,
1849
+ "step": 5140
1850
+ },
1851
+ {
1852
+ "epoch": 1.6688227684346701,
1853
+ "grad_norm": 0.3626604378223419,
1854
+ "learning_rate": 3.6066150598170304e-05,
1855
+ "loss": 0.3319,
1856
+ "step": 5160
1857
+ },
1858
+ {
1859
+ "epoch": 1.6752910737386806,
1860
+ "grad_norm": 0.39335256814956665,
1861
+ "learning_rate": 3.536242083040113e-05,
1862
+ "loss": 0.3234,
1863
+ "step": 5180
1864
+ },
1865
+ {
1866
+ "epoch": 1.6817593790426908,
1867
+ "grad_norm": 0.35490670800209045,
1868
+ "learning_rate": 3.465869106263195e-05,
1869
+ "loss": 0.3357,
1870
+ "step": 5200
1871
+ },
1872
+ {
1873
+ "epoch": 1.688227684346701,
1874
+ "grad_norm": 0.303774356842041,
1875
+ "learning_rate": 3.3954961294862776e-05,
1876
+ "loss": 0.3609,
1877
+ "step": 5220
1878
+ },
1879
+ {
1880
+ "epoch": 1.6946959896507114,
1881
+ "grad_norm": 0.3438192307949066,
1882
+ "learning_rate": 3.3251231527093596e-05,
1883
+ "loss": 0.3351,
1884
+ "step": 5240
1885
+ },
1886
+ {
1887
+ "epoch": 1.7011642949547219,
1888
+ "grad_norm": 0.3261934518814087,
1889
+ "learning_rate": 3.254750175932442e-05,
1890
+ "loss": 0.3192,
1891
+ "step": 5260
1892
+ },
1893
+ {
1894
+ "epoch": 1.7076326002587323,
1895
+ "grad_norm": 0.39429745078086853,
1896
+ "learning_rate": 3.184377199155524e-05,
1897
+ "loss": 0.3479,
1898
+ "step": 5280
1899
+ },
1900
+ {
1901
+ "epoch": 1.7141009055627425,
1902
+ "grad_norm": 0.4267764985561371,
1903
+ "learning_rate": 3.114004222378607e-05,
1904
+ "loss": 0.3306,
1905
+ "step": 5300
1906
+ },
1907
+ {
1908
+ "epoch": 1.720569210866753,
1909
+ "grad_norm": 0.3425467908382416,
1910
+ "learning_rate": 3.0436312456016892e-05,
1911
+ "loss": 0.3356,
1912
+ "step": 5320
1913
+ },
1914
+ {
1915
+ "epoch": 1.7270375161707632,
1916
+ "grad_norm": 0.3064535856246948,
1917
+ "learning_rate": 2.9732582688247712e-05,
1918
+ "loss": 0.3341,
1919
+ "step": 5340
1920
+ },
1921
+ {
1922
+ "epoch": 1.7335058214747736,
1923
+ "grad_norm": 0.3956790566444397,
1924
+ "learning_rate": 2.902885292047854e-05,
1925
+ "loss": 0.3375,
1926
+ "step": 5360
1927
+ },
1928
+ {
1929
+ "epoch": 1.739974126778784,
1930
+ "grad_norm": 0.36128687858581543,
1931
+ "learning_rate": 2.8325123152709358e-05,
1932
+ "loss": 0.3395,
1933
+ "step": 5380
1934
+ },
1935
+ {
1936
+ "epoch": 1.7464424320827943,
1937
+ "grad_norm": 0.33011215925216675,
1938
+ "learning_rate": 2.7621393384940185e-05,
1939
+ "loss": 0.3078,
1940
+ "step": 5400
1941
+ },
1942
+ {
1943
+ "epoch": 1.7529107373868047,
1944
+ "grad_norm": 0.3741130530834198,
1945
+ "learning_rate": 2.6917663617171008e-05,
1946
+ "loss": 0.362,
1947
+ "step": 5420
1948
+ },
1949
+ {
1950
+ "epoch": 1.759379042690815,
1951
+ "grad_norm": 0.319671094417572,
1952
+ "learning_rate": 2.6213933849401827e-05,
1953
+ "loss": 0.3105,
1954
+ "step": 5440
1955
+ },
1956
+ {
1957
+ "epoch": 1.7658473479948253,
1958
+ "grad_norm": 0.4425293505191803,
1959
+ "learning_rate": 2.5510204081632654e-05,
1960
+ "loss": 0.3327,
1961
+ "step": 5460
1962
+ },
1963
+ {
1964
+ "epoch": 1.7723156532988358,
1965
+ "grad_norm": 0.43660980463027954,
1966
+ "learning_rate": 2.4806474313863477e-05,
1967
+ "loss": 0.3057,
1968
+ "step": 5480
1969
+ },
1970
+ {
1971
+ "epoch": 1.7787839586028462,
1972
+ "grad_norm": 0.3476182222366333,
1973
+ "learning_rate": 2.41027445460943e-05,
1974
+ "loss": 0.3362,
1975
+ "step": 5500
1976
+ },
1977
+ {
1978
+ "epoch": 1.7852522639068564,
1979
+ "grad_norm": 0.37895113229751587,
1980
+ "learning_rate": 2.3399014778325123e-05,
1981
+ "loss": 0.337,
1982
+ "step": 5520
1983
+ },
1984
+ {
1985
+ "epoch": 1.7917205692108666,
1986
+ "grad_norm": 0.38524091243743896,
1987
+ "learning_rate": 2.2695285010555947e-05,
1988
+ "loss": 0.362,
1989
+ "step": 5540
1990
+ },
1991
+ {
1992
+ "epoch": 1.798188874514877,
1993
+ "grad_norm": 0.38090500235557556,
1994
+ "learning_rate": 2.199155524278677e-05,
1995
+ "loss": 0.3268,
1996
+ "step": 5560
1997
+ },
1998
+ {
1999
+ "epoch": 1.8046571798188875,
2000
+ "grad_norm": 0.4094742238521576,
2001
+ "learning_rate": 2.1287825475017596e-05,
2002
+ "loss": 0.3353,
2003
+ "step": 5580
2004
+ },
2005
+ {
2006
+ "epoch": 1.811125485122898,
2007
+ "grad_norm": 0.37963709235191345,
2008
+ "learning_rate": 2.058409570724842e-05,
2009
+ "loss": 0.3197,
2010
+ "step": 5600
2011
+ },
2012
+ {
2013
+ "epoch": 1.8175937904269082,
2014
+ "grad_norm": 0.4362301230430603,
2015
+ "learning_rate": 1.988036593947924e-05,
2016
+ "loss": 0.3405,
2017
+ "step": 5620
2018
+ },
2019
+ {
2020
+ "epoch": 1.8240620957309184,
2021
+ "grad_norm": 0.3662102520465851,
2022
+ "learning_rate": 1.9176636171710062e-05,
2023
+ "loss": 0.3262,
2024
+ "step": 5640
2025
+ },
2026
+ {
2027
+ "epoch": 1.8305304010349288,
2028
+ "grad_norm": 0.3674371838569641,
2029
+ "learning_rate": 1.847290640394089e-05,
2030
+ "loss": 0.3394,
2031
+ "step": 5660
2032
+ },
2033
+ {
2034
+ "epoch": 1.8369987063389392,
2035
+ "grad_norm": 0.31638914346694946,
2036
+ "learning_rate": 1.7769176636171712e-05,
2037
+ "loss": 0.3072,
2038
+ "step": 5680
2039
+ },
2040
+ {
2041
+ "epoch": 1.8434670116429497,
2042
+ "grad_norm": 0.39859539270401,
2043
+ "learning_rate": 1.7065446868402535e-05,
2044
+ "loss": 0.3398,
2045
+ "step": 5700
2046
+ },
2047
+ {
2048
+ "epoch": 1.84993531694696,
2049
+ "grad_norm": 0.37528204917907715,
2050
+ "learning_rate": 1.6361717100633358e-05,
2051
+ "loss": 0.3362,
2052
+ "step": 5720
2053
+ },
2054
+ {
2055
+ "epoch": 1.85640362225097,
2056
+ "grad_norm": 0.40589115023612976,
2057
+ "learning_rate": 1.565798733286418e-05,
2058
+ "loss": 0.3369,
2059
+ "step": 5740
2060
+ },
2061
+ {
2062
+ "epoch": 1.8628719275549805,
2063
+ "grad_norm": 0.3010273277759552,
2064
+ "learning_rate": 1.4954257565095003e-05,
2065
+ "loss": 0.3293,
2066
+ "step": 5760
2067
+ },
2068
+ {
2069
+ "epoch": 1.869340232858991,
2070
+ "grad_norm": 0.47433990240097046,
2071
+ "learning_rate": 1.4250527797325827e-05,
2072
+ "loss": 0.3137,
2073
+ "step": 5780
2074
+ },
2075
+ {
2076
+ "epoch": 1.8758085381630014,
2077
+ "grad_norm": 0.3495091199874878,
2078
+ "learning_rate": 1.354679802955665e-05,
2079
+ "loss": 0.3077,
2080
+ "step": 5800
2081
+ },
2082
+ {
2083
+ "epoch": 1.8822768434670116,
2084
+ "grad_norm": 0.33797210454940796,
2085
+ "learning_rate": 1.2843068261787475e-05,
2086
+ "loss": 0.3454,
2087
+ "step": 5820
2088
+ },
2089
+ {
2090
+ "epoch": 1.8887451487710218,
2091
+ "grad_norm": 0.34069564938545227,
2092
+ "learning_rate": 1.2139338494018297e-05,
2093
+ "loss": 0.3087,
2094
+ "step": 5840
2095
+ },
2096
+ {
2097
+ "epoch": 1.8952134540750323,
2098
+ "grad_norm": 0.4089032709598541,
2099
+ "learning_rate": 1.1435608726249122e-05,
2100
+ "loss": 0.3487,
2101
+ "step": 5860
2102
+ },
2103
+ {
2104
+ "epoch": 1.9016817593790427,
2105
+ "grad_norm": 0.3203008472919464,
2106
+ "learning_rate": 1.0731878958479945e-05,
2107
+ "loss": 0.3301,
2108
+ "step": 5880
2109
+ },
2110
+ {
2111
+ "epoch": 1.9081500646830531,
2112
+ "grad_norm": 0.3505829870700836,
2113
+ "learning_rate": 1.0028149190710768e-05,
2114
+ "loss": 0.3319,
2115
+ "step": 5900
2116
+ },
2117
+ {
2118
+ "epoch": 1.9146183699870634,
2119
+ "grad_norm": 0.29925537109375,
2120
+ "learning_rate": 9.324419422941591e-06,
2121
+ "loss": 0.3317,
2122
+ "step": 5920
2123
+ },
2124
+ {
2125
+ "epoch": 1.9210866752910736,
2126
+ "grad_norm": 0.4060503840446472,
2127
+ "learning_rate": 8.620689655172414e-06,
2128
+ "loss": 0.3393,
2129
+ "step": 5940
2130
+ },
2131
+ {
2132
+ "epoch": 1.927554980595084,
2133
+ "grad_norm": 0.4205261170864105,
2134
+ "learning_rate": 7.916959887403237e-06,
2135
+ "loss": 0.3375,
2136
+ "step": 5960
2137
+ },
2138
+ {
2139
+ "epoch": 1.9340232858990944,
2140
+ "grad_norm": 0.44757845997810364,
2141
+ "learning_rate": 7.213230119634061e-06,
2142
+ "loss": 0.3329,
2143
+ "step": 5980
2144
+ },
2145
+ {
2146
+ "epoch": 1.9404915912031049,
2147
+ "grad_norm": 0.5326984524726868,
2148
+ "learning_rate": 6.509500351864884e-06,
2149
+ "loss": 0.342,
2150
+ "step": 6000
2151
+ },
2152
+ {
2153
+ "epoch": 1.9404915912031049,
2154
+ "eval_loss": 0.5839738845825195,
2155
+ "eval_runtime": 967.7859,
2156
+ "eval_samples_per_second": 6.331,
2157
+ "eval_steps_per_second": 0.791,
2158
+ "step": 6000
2159
+ }
2160
+ ],
2161
+ "logging_steps": 20,
2162
+ "max_steps": 6184,
2163
+ "num_input_tokens_seen": 0,
2164
+ "num_train_epochs": 2,
2165
+ "save_steps": 1000,
2166
+ "stateful_callbacks": {
2167
+ "TrainerControl": {
2168
+ "args": {
2169
+ "should_epoch_stop": false,
2170
+ "should_evaluate": false,
2171
+ "should_log": false,
2172
+ "should_save": true,
2173
+ "should_training_stop": false
2174
+ },
2175
+ "attributes": {}
2176
+ }
2177
+ },
2178
+ "total_flos": 7.991147541685469e+18,
2179
+ "train_batch_size": 8,
2180
+ "trial_name": null,
2181
+ "trial_params": null
2182
+ }
ultival_mistral/checkpoint-6000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8e5db938f70ff2ddfe787ade649d14bb55cc2be7a65bd63a20c3303bf68492
3
+ size 5368
ultival_mistral/checkpoint-6184/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Ministral-8B-Instruct-2410
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
ultival_mistral/checkpoint-6184/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Ministral-8B-Instruct-2410",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 32,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "o_proj",
28
+ "k_proj",
29
+ "down_proj",
30
+ "q_proj",
31
+ "gate_proj",
32
+ "up_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
ultival_mistral/checkpoint-6184/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cdcd01ee05b36166eb8d36352e3dd9de01a098cc56f647bf981c0fc5ec37393
3
+ size 349243752
ultival_mistral/checkpoint-6184/chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{% if messages[1]['role'] == 'user' %}{{ '[INST] ' + messages[0]['content'] + ' ' + messages[1]['content'] + ' [/INST]' }}{% set loop_messages = messages[2:] %}{% else %}{{ '[INST] ' + messages[0]['content'] + ' [/INST]' }}{% set loop_messages = messages[1:] %}{% endif %}{% else %}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}
ultival_mistral/checkpoint-6184/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52eac4bb89fa48810acf94bc308f7d720f59643d37ef36715604087b81444897
3
+ size 177919028
ultival_mistral/checkpoint-6184/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf932362a907148c2c99f7826e21fa7280b5016d990a6358e60ec3cec98b016
3
+ size 14244
ultival_mistral/checkpoint-6184/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd069de76fd3803039b7454910298a611b2a1a5486b5af24595ccb064fdf7323
3
+ size 1064
ultival_mistral/checkpoint-6184/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<pad>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
ultival_mistral/checkpoint-6184/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be2965d3e435dbdef9376e5bdb3dcc2bfc7e271aa5152ae896cf53c7e920adf8
3
+ size 17078235
ultival_mistral/checkpoint-6184/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
ultival_mistral/checkpoint-6184/trainer_state.json ADDED
@@ -0,0 +1,2245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 6000,
3
+ "best_metric": 0.5839738845825195,
4
+ "best_model_checkpoint": "./ultival_run2/checkpoint-6000",
5
+ "epoch": 2.0,
6
+ "eval_steps": 1000,
7
+ "global_step": 6184,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00646830530401035,
14
+ "grad_norm": 0.8976479172706604,
15
+ "learning_rate": 7.6e-06,
16
+ "loss": 1.5544,
17
+ "step": 20
18
+ },
19
+ {
20
+ "epoch": 0.0129366106080207,
21
+ "grad_norm": 0.726784348487854,
22
+ "learning_rate": 1.56e-05,
23
+ "loss": 1.242,
24
+ "step": 40
25
+ },
26
+ {
27
+ "epoch": 0.019404915912031046,
28
+ "grad_norm": 0.4917074143886566,
29
+ "learning_rate": 2.36e-05,
30
+ "loss": 0.6694,
31
+ "step": 60
32
+ },
33
+ {
34
+ "epoch": 0.0258732212160414,
35
+ "grad_norm": 0.3471176326274872,
36
+ "learning_rate": 3.16e-05,
37
+ "loss": 0.5334,
38
+ "step": 80
39
+ },
40
+ {
41
+ "epoch": 0.03234152652005175,
42
+ "grad_norm": 0.41759616136550903,
43
+ "learning_rate": 3.960000000000001e-05,
44
+ "loss": 0.5435,
45
+ "step": 100
46
+ },
47
+ {
48
+ "epoch": 0.03880983182406209,
49
+ "grad_norm": 0.3755040466785431,
50
+ "learning_rate": 4.76e-05,
51
+ "loss": 0.5684,
52
+ "step": 120
53
+ },
54
+ {
55
+ "epoch": 0.045278137128072445,
56
+ "grad_norm": 0.37058988213539124,
57
+ "learning_rate": 5.560000000000001e-05,
58
+ "loss": 0.5329,
59
+ "step": 140
60
+ },
61
+ {
62
+ "epoch": 0.0517464424320828,
63
+ "grad_norm": 0.42603132128715515,
64
+ "learning_rate": 6.36e-05,
65
+ "loss": 0.5457,
66
+ "step": 160
67
+ },
68
+ {
69
+ "epoch": 0.05821474773609314,
70
+ "grad_norm": 0.3154866099357605,
71
+ "learning_rate": 7.16e-05,
72
+ "loss": 0.5472,
73
+ "step": 180
74
+ },
75
+ {
76
+ "epoch": 0.0646830530401035,
77
+ "grad_norm": 0.3919835388660431,
78
+ "learning_rate": 7.960000000000001e-05,
79
+ "loss": 0.4675,
80
+ "step": 200
81
+ },
82
+ {
83
+ "epoch": 0.07115135834411385,
84
+ "grad_norm": 0.32180657982826233,
85
+ "learning_rate": 8.76e-05,
86
+ "loss": 0.5078,
87
+ "step": 220
88
+ },
89
+ {
90
+ "epoch": 0.07761966364812418,
91
+ "grad_norm": 0.3246982991695404,
92
+ "learning_rate": 9.56e-05,
93
+ "loss": 0.488,
94
+ "step": 240
95
+ },
96
+ {
97
+ "epoch": 0.08408796895213454,
98
+ "grad_norm": 0.341487318277359,
99
+ "learning_rate": 0.00010360000000000001,
100
+ "loss": 0.5058,
101
+ "step": 260
102
+ },
103
+ {
104
+ "epoch": 0.09055627425614489,
105
+ "grad_norm": 0.3597952127456665,
106
+ "learning_rate": 0.00011160000000000002,
107
+ "loss": 0.4574,
108
+ "step": 280
109
+ },
110
+ {
111
+ "epoch": 0.09702457956015524,
112
+ "grad_norm": 0.40912091732025146,
113
+ "learning_rate": 0.00011960000000000001,
114
+ "loss": 0.4931,
115
+ "step": 300
116
+ },
117
+ {
118
+ "epoch": 0.1034928848641656,
119
+ "grad_norm": 0.37430697679519653,
120
+ "learning_rate": 0.0001276,
121
+ "loss": 0.4724,
122
+ "step": 320
123
+ },
124
+ {
125
+ "epoch": 0.10996119016817593,
126
+ "grad_norm": 0.3920972943305969,
127
+ "learning_rate": 0.00013560000000000002,
128
+ "loss": 0.4997,
129
+ "step": 340
130
+ },
131
+ {
132
+ "epoch": 0.11642949547218628,
133
+ "grad_norm": 0.32511982321739197,
134
+ "learning_rate": 0.0001436,
135
+ "loss": 0.4889,
136
+ "step": 360
137
+ },
138
+ {
139
+ "epoch": 0.12289780077619664,
140
+ "grad_norm": 0.3926583528518677,
141
+ "learning_rate": 0.0001516,
142
+ "loss": 0.4772,
143
+ "step": 380
144
+ },
145
+ {
146
+ "epoch": 0.129366106080207,
147
+ "grad_norm": 0.3535909950733185,
148
+ "learning_rate": 0.0001596,
149
+ "loss": 0.465,
150
+ "step": 400
151
+ },
152
+ {
153
+ "epoch": 0.13583441138421734,
154
+ "grad_norm": 0.34381625056266785,
155
+ "learning_rate": 0.0001676,
156
+ "loss": 0.4934,
157
+ "step": 420
158
+ },
159
+ {
160
+ "epoch": 0.1423027166882277,
161
+ "grad_norm": 0.3753703236579895,
162
+ "learning_rate": 0.0001756,
163
+ "loss": 0.5068,
164
+ "step": 440
165
+ },
166
+ {
167
+ "epoch": 0.14877102199223805,
168
+ "grad_norm": 0.34895479679107666,
169
+ "learning_rate": 0.00018360000000000002,
170
+ "loss": 0.5186,
171
+ "step": 460
172
+ },
173
+ {
174
+ "epoch": 0.15523932729624837,
175
+ "grad_norm": 0.24720332026481628,
176
+ "learning_rate": 0.0001916,
177
+ "loss": 0.461,
178
+ "step": 480
179
+ },
180
+ {
181
+ "epoch": 0.16170763260025872,
182
+ "grad_norm": 0.3713696002960205,
183
+ "learning_rate": 0.0001996,
184
+ "loss": 0.4932,
185
+ "step": 500
186
+ },
187
+ {
188
+ "epoch": 0.16817593790426907,
189
+ "grad_norm": 0.3724994659423828,
190
+ "learning_rate": 0.00019933145672061927,
191
+ "loss": 0.5036,
192
+ "step": 520
193
+ },
194
+ {
195
+ "epoch": 0.17464424320827943,
196
+ "grad_norm": 0.3562956750392914,
197
+ "learning_rate": 0.00019862772695285012,
198
+ "loss": 0.4871,
199
+ "step": 540
200
+ },
201
+ {
202
+ "epoch": 0.18111254851228978,
203
+ "grad_norm": 0.331737756729126,
204
+ "learning_rate": 0.00019792399718508095,
205
+ "loss": 0.5178,
206
+ "step": 560
207
+ },
208
+ {
209
+ "epoch": 0.18758085381630013,
210
+ "grad_norm": 0.34049859642982483,
211
+ "learning_rate": 0.00019722026741731178,
212
+ "loss": 0.488,
213
+ "step": 580
214
+ },
215
+ {
216
+ "epoch": 0.19404915912031048,
217
+ "grad_norm": 0.36076661944389343,
218
+ "learning_rate": 0.00019651653764954258,
219
+ "loss": 0.4821,
220
+ "step": 600
221
+ },
222
+ {
223
+ "epoch": 0.20051746442432083,
224
+ "grad_norm": 0.3059399127960205,
225
+ "learning_rate": 0.0001958128078817734,
226
+ "loss": 0.4924,
227
+ "step": 620
228
+ },
229
+ {
230
+ "epoch": 0.2069857697283312,
231
+ "grad_norm": 0.30692920088768005,
232
+ "learning_rate": 0.00019510907811400423,
233
+ "loss": 0.4822,
234
+ "step": 640
235
+ },
236
+ {
237
+ "epoch": 0.21345407503234154,
238
+ "grad_norm": 0.3319448232650757,
239
+ "learning_rate": 0.00019440534834623506,
240
+ "loss": 0.4607,
241
+ "step": 660
242
+ },
243
+ {
244
+ "epoch": 0.21992238033635186,
245
+ "grad_norm": 0.3047769069671631,
246
+ "learning_rate": 0.00019370161857846588,
247
+ "loss": 0.5357,
248
+ "step": 680
249
+ },
250
+ {
251
+ "epoch": 0.22639068564036222,
252
+ "grad_norm": 0.4353258013725281,
253
+ "learning_rate": 0.00019299788881069668,
254
+ "loss": 0.4738,
255
+ "step": 700
256
+ },
257
+ {
258
+ "epoch": 0.23285899094437257,
259
+ "grad_norm": 0.30877336859703064,
260
+ "learning_rate": 0.00019229415904292754,
261
+ "loss": 0.5128,
262
+ "step": 720
263
+ },
264
+ {
265
+ "epoch": 0.23932729624838292,
266
+ "grad_norm": 0.3188530206680298,
267
+ "learning_rate": 0.00019159042927515836,
268
+ "loss": 0.4651,
269
+ "step": 740
270
+ },
271
+ {
272
+ "epoch": 0.24579560155239327,
273
+ "grad_norm": 0.40678778290748596,
274
+ "learning_rate": 0.0001908866995073892,
275
+ "loss": 0.5112,
276
+ "step": 760
277
+ },
278
+ {
279
+ "epoch": 0.2522639068564036,
280
+ "grad_norm": 0.34188327193260193,
281
+ "learning_rate": 0.00019018296973962,
282
+ "loss": 0.487,
283
+ "step": 780
284
+ },
285
+ {
286
+ "epoch": 0.258732212160414,
287
+ "grad_norm": 0.3002854585647583,
288
+ "learning_rate": 0.00018947923997185081,
289
+ "loss": 0.512,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.2652005174644243,
294
+ "grad_norm": 0.35756272077560425,
295
+ "learning_rate": 0.00018877551020408164,
296
+ "loss": 0.4812,
297
+ "step": 820
298
+ },
299
+ {
300
+ "epoch": 0.2716688227684347,
301
+ "grad_norm": 0.33527258038520813,
302
+ "learning_rate": 0.00018807178043631247,
303
+ "loss": 0.4895,
304
+ "step": 840
305
+ },
306
+ {
307
+ "epoch": 0.278137128072445,
308
+ "grad_norm": 0.3472883105278015,
309
+ "learning_rate": 0.0001873680506685433,
310
+ "loss": 0.4671,
311
+ "step": 860
312
+ },
313
+ {
314
+ "epoch": 0.2846054333764554,
315
+ "grad_norm": 0.3604671061038971,
316
+ "learning_rate": 0.0001866643209007741,
317
+ "loss": 0.5233,
318
+ "step": 880
319
+ },
320
+ {
321
+ "epoch": 0.2910737386804657,
322
+ "grad_norm": 0.31021153926849365,
323
+ "learning_rate": 0.00018596059113300492,
324
+ "loss": 0.491,
325
+ "step": 900
326
+ },
327
+ {
328
+ "epoch": 0.2975420439844761,
329
+ "grad_norm": 0.6417067050933838,
330
+ "learning_rate": 0.00018525686136523577,
331
+ "loss": 0.4858,
332
+ "step": 920
333
+ },
334
+ {
335
+ "epoch": 0.3040103492884864,
336
+ "grad_norm": 0.33719781041145325,
337
+ "learning_rate": 0.0001845531315974666,
338
+ "loss": 0.4722,
339
+ "step": 940
340
+ },
341
+ {
342
+ "epoch": 0.31047865459249674,
343
+ "grad_norm": 0.3568169176578522,
344
+ "learning_rate": 0.0001838494018296974,
345
+ "loss": 0.5088,
346
+ "step": 960
347
+ },
348
+ {
349
+ "epoch": 0.3169469598965071,
350
+ "grad_norm": 0.42162615060806274,
351
+ "learning_rate": 0.00018314567206192823,
352
+ "loss": 0.5091,
353
+ "step": 980
354
+ },
355
+ {
356
+ "epoch": 0.32341526520051744,
357
+ "grad_norm": 0.3212260603904724,
358
+ "learning_rate": 0.00018244194229415905,
359
+ "loss": 0.4654,
360
+ "step": 1000
361
+ },
362
+ {
363
+ "epoch": 0.32341526520051744,
364
+ "eval_loss": 0.6167479753494263,
365
+ "eval_runtime": 1013.3659,
366
+ "eval_samples_per_second": 6.046,
367
+ "eval_steps_per_second": 0.756,
368
+ "step": 1000
369
+ },
370
+ {
371
+ "epoch": 0.3298835705045278,
372
+ "grad_norm": 0.3184376060962677,
373
+ "learning_rate": 0.00018173821252638988,
374
+ "loss": 0.4712,
375
+ "step": 1020
376
+ },
377
+ {
378
+ "epoch": 0.33635187580853815,
379
+ "grad_norm": 0.3114645183086395,
380
+ "learning_rate": 0.0001810344827586207,
381
+ "loss": 0.4393,
382
+ "step": 1040
383
+ },
384
+ {
385
+ "epoch": 0.3428201811125485,
386
+ "grad_norm": 0.30122488737106323,
387
+ "learning_rate": 0.0001803307529908515,
388
+ "loss": 0.4882,
389
+ "step": 1060
390
+ },
391
+ {
392
+ "epoch": 0.34928848641655885,
393
+ "grad_norm": 0.2877393066883087,
394
+ "learning_rate": 0.00017962702322308233,
395
+ "loss": 0.413,
396
+ "step": 1080
397
+ },
398
+ {
399
+ "epoch": 0.35575679172056923,
400
+ "grad_norm": 0.33227887749671936,
401
+ "learning_rate": 0.00017892329345531319,
402
+ "loss": 0.4527,
403
+ "step": 1100
404
+ },
405
+ {
406
+ "epoch": 0.36222509702457956,
407
+ "grad_norm": 0.33423128724098206,
408
+ "learning_rate": 0.000178219563687544,
409
+ "loss": 0.4604,
410
+ "step": 1120
411
+ },
412
+ {
413
+ "epoch": 0.36869340232858994,
414
+ "grad_norm": 0.36922487616539,
415
+ "learning_rate": 0.0001775158339197748,
416
+ "loss": 0.5044,
417
+ "step": 1140
418
+ },
419
+ {
420
+ "epoch": 0.37516170763260026,
421
+ "grad_norm": 0.34816262125968933,
422
+ "learning_rate": 0.00017681210415200564,
423
+ "loss": 0.4898,
424
+ "step": 1160
425
+ },
426
+ {
427
+ "epoch": 0.3816300129366106,
428
+ "grad_norm": 0.32086044549942017,
429
+ "learning_rate": 0.00017610837438423646,
430
+ "loss": 0.4647,
431
+ "step": 1180
432
+ },
433
+ {
434
+ "epoch": 0.38809831824062097,
435
+ "grad_norm": 0.3879236876964569,
436
+ "learning_rate": 0.0001754046446164673,
437
+ "loss": 0.5179,
438
+ "step": 1200
439
+ },
440
+ {
441
+ "epoch": 0.3945666235446313,
442
+ "grad_norm": 0.2899722158908844,
443
+ "learning_rate": 0.00017470091484869812,
444
+ "loss": 0.4786,
445
+ "step": 1220
446
+ },
447
+ {
448
+ "epoch": 0.40103492884864167,
449
+ "grad_norm": 0.3844458758831024,
450
+ "learning_rate": 0.00017399718508092892,
451
+ "loss": 0.4513,
452
+ "step": 1240
453
+ },
454
+ {
455
+ "epoch": 0.407503234152652,
456
+ "grad_norm": 0.3678169548511505,
457
+ "learning_rate": 0.00017329345531315974,
458
+ "loss": 0.4789,
459
+ "step": 1260
460
+ },
461
+ {
462
+ "epoch": 0.4139715394566624,
463
+ "grad_norm": 0.37138116359710693,
464
+ "learning_rate": 0.00017258972554539057,
465
+ "loss": 0.4694,
466
+ "step": 1280
467
+ },
468
+ {
469
+ "epoch": 0.4204398447606727,
470
+ "grad_norm": 0.3604221045970917,
471
+ "learning_rate": 0.00017188599577762142,
472
+ "loss": 0.4496,
473
+ "step": 1300
474
+ },
475
+ {
476
+ "epoch": 0.4269081500646831,
477
+ "grad_norm": 0.33961811661720276,
478
+ "learning_rate": 0.00017118226600985222,
479
+ "loss": 0.4696,
480
+ "step": 1320
481
+ },
482
+ {
483
+ "epoch": 0.4333764553686934,
484
+ "grad_norm": 0.3244430124759674,
485
+ "learning_rate": 0.00017047853624208305,
486
+ "loss": 0.4322,
487
+ "step": 1340
488
+ },
489
+ {
490
+ "epoch": 0.4398447606727037,
491
+ "grad_norm": 0.2879267930984497,
492
+ "learning_rate": 0.00016977480647431388,
493
+ "loss": 0.4646,
494
+ "step": 1360
495
+ },
496
+ {
497
+ "epoch": 0.4463130659767141,
498
+ "grad_norm": 0.3437345027923584,
499
+ "learning_rate": 0.0001690710767065447,
500
+ "loss": 0.4711,
501
+ "step": 1380
502
+ },
503
+ {
504
+ "epoch": 0.45278137128072443,
505
+ "grad_norm": 0.3166882395744324,
506
+ "learning_rate": 0.00016836734693877553,
507
+ "loss": 0.486,
508
+ "step": 1400
509
+ },
510
+ {
511
+ "epoch": 0.4592496765847348,
512
+ "grad_norm": 0.3677555322647095,
513
+ "learning_rate": 0.00016766361717100633,
514
+ "loss": 0.4677,
515
+ "step": 1420
516
+ },
517
+ {
518
+ "epoch": 0.46571798188874514,
519
+ "grad_norm": 0.32657769322395325,
520
+ "learning_rate": 0.00016695988740323715,
521
+ "loss": 0.4649,
522
+ "step": 1440
523
+ },
524
+ {
525
+ "epoch": 0.4721862871927555,
526
+ "grad_norm": 0.2980731427669525,
527
+ "learning_rate": 0.00016625615763546798,
528
+ "loss": 0.4547,
529
+ "step": 1460
530
+ },
531
+ {
532
+ "epoch": 0.47865459249676584,
533
+ "grad_norm": 0.3690549433231354,
534
+ "learning_rate": 0.0001655524278676988,
535
+ "loss": 0.4512,
536
+ "step": 1480
537
+ },
538
+ {
539
+ "epoch": 0.4851228978007762,
540
+ "grad_norm": 0.31456947326660156,
541
+ "learning_rate": 0.00016484869809992963,
542
+ "loss": 0.4675,
543
+ "step": 1500
544
+ },
545
+ {
546
+ "epoch": 0.49159120310478654,
547
+ "grad_norm": 0.3823802173137665,
548
+ "learning_rate": 0.00016414496833216046,
549
+ "loss": 0.4849,
550
+ "step": 1520
551
+ },
552
+ {
553
+ "epoch": 0.49805950840879687,
554
+ "grad_norm": 0.33280324935913086,
555
+ "learning_rate": 0.0001634412385643913,
556
+ "loss": 0.4876,
557
+ "step": 1540
558
+ },
559
+ {
560
+ "epoch": 0.5045278137128072,
561
+ "grad_norm": 0.28788062930107117,
562
+ "learning_rate": 0.00016273750879662211,
563
+ "loss": 0.426,
564
+ "step": 1560
565
+ },
566
+ {
567
+ "epoch": 0.5109961190168176,
568
+ "grad_norm": 0.3809375762939453,
569
+ "learning_rate": 0.00016203377902885294,
570
+ "loss": 0.4676,
571
+ "step": 1580
572
+ },
573
+ {
574
+ "epoch": 0.517464424320828,
575
+ "grad_norm": 0.27940312027931213,
576
+ "learning_rate": 0.00016133004926108374,
577
+ "loss": 0.4413,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 0.5239327296248383,
582
+ "grad_norm": 0.30946919322013855,
583
+ "learning_rate": 0.00016062631949331457,
584
+ "loss": 0.4395,
585
+ "step": 1620
586
+ },
587
+ {
588
+ "epoch": 0.5304010349288486,
589
+ "grad_norm": 0.36931583285331726,
590
+ "learning_rate": 0.0001599225897255454,
591
+ "loss": 0.4565,
592
+ "step": 1640
593
+ },
594
+ {
595
+ "epoch": 0.536869340232859,
596
+ "grad_norm": 0.3652470111846924,
597
+ "learning_rate": 0.00015921885995777622,
598
+ "loss": 0.4522,
599
+ "step": 1660
600
+ },
601
+ {
602
+ "epoch": 0.5433376455368694,
603
+ "grad_norm": 0.3060782551765442,
604
+ "learning_rate": 0.00015851513019000705,
605
+ "loss": 0.45,
606
+ "step": 1680
607
+ },
608
+ {
609
+ "epoch": 0.5498059508408797,
610
+ "grad_norm": 0.3275803029537201,
611
+ "learning_rate": 0.00015781140042223787,
612
+ "loss": 0.445,
613
+ "step": 1700
614
+ },
615
+ {
616
+ "epoch": 0.55627425614489,
617
+ "grad_norm": 0.34865352511405945,
618
+ "learning_rate": 0.0001571076706544687,
619
+ "loss": 0.4306,
620
+ "step": 1720
621
+ },
622
+ {
623
+ "epoch": 0.5627425614489003,
624
+ "grad_norm": 0.42525383830070496,
625
+ "learning_rate": 0.00015640394088669953,
626
+ "loss": 0.4544,
627
+ "step": 1740
628
+ },
629
+ {
630
+ "epoch": 0.5692108667529108,
631
+ "grad_norm": 0.32165050506591797,
632
+ "learning_rate": 0.00015570021111893035,
633
+ "loss": 0.484,
634
+ "step": 1760
635
+ },
636
+ {
637
+ "epoch": 0.5756791720569211,
638
+ "grad_norm": 0.3213440477848053,
639
+ "learning_rate": 0.00015499648135116115,
640
+ "loss": 0.4469,
641
+ "step": 1780
642
+ },
643
+ {
644
+ "epoch": 0.5821474773609314,
645
+ "grad_norm": 0.42835432291030884,
646
+ "learning_rate": 0.00015429275158339198,
647
+ "loss": 0.4592,
648
+ "step": 1800
649
+ },
650
+ {
651
+ "epoch": 0.5886157826649417,
652
+ "grad_norm": 0.3861497938632965,
653
+ "learning_rate": 0.0001535890218156228,
654
+ "loss": 0.4607,
655
+ "step": 1820
656
+ },
657
+ {
658
+ "epoch": 0.5950840879689522,
659
+ "grad_norm": 0.4156148135662079,
660
+ "learning_rate": 0.00015288529204785363,
661
+ "loss": 0.4541,
662
+ "step": 1840
663
+ },
664
+ {
665
+ "epoch": 0.6015523932729625,
666
+ "grad_norm": 0.32371824979782104,
667
+ "learning_rate": 0.00015218156228008446,
668
+ "loss": 0.4287,
669
+ "step": 1860
670
+ },
671
+ {
672
+ "epoch": 0.6080206985769728,
673
+ "grad_norm": 0.363788366317749,
674
+ "learning_rate": 0.00015147783251231528,
675
+ "loss": 0.428,
676
+ "step": 1880
677
+ },
678
+ {
679
+ "epoch": 0.6144890038809832,
680
+ "grad_norm": 0.3932839334011078,
681
+ "learning_rate": 0.0001507741027445461,
682
+ "loss": 0.4359,
683
+ "step": 1900
684
+ },
685
+ {
686
+ "epoch": 0.6209573091849935,
687
+ "grad_norm": 0.3009307086467743,
688
+ "learning_rate": 0.00015007037297677694,
689
+ "loss": 0.3976,
690
+ "step": 1920
691
+ },
692
+ {
693
+ "epoch": 0.6274256144890039,
694
+ "grad_norm": 0.3484659194946289,
695
+ "learning_rate": 0.00014936664320900776,
696
+ "loss": 0.4584,
697
+ "step": 1940
698
+ },
699
+ {
700
+ "epoch": 0.6338939197930142,
701
+ "grad_norm": 0.33728504180908203,
702
+ "learning_rate": 0.00014866291344123856,
703
+ "loss": 0.4543,
704
+ "step": 1960
705
+ },
706
+ {
707
+ "epoch": 0.6403622250970246,
708
+ "grad_norm": 0.3207024037837982,
709
+ "learning_rate": 0.0001479591836734694,
710
+ "loss": 0.4611,
711
+ "step": 1980
712
+ },
713
+ {
714
+ "epoch": 0.6468305304010349,
715
+ "grad_norm": 0.29134076833724976,
716
+ "learning_rate": 0.00014725545390570022,
717
+ "loss": 0.439,
718
+ "step": 2000
719
+ },
720
+ {
721
+ "epoch": 0.6468305304010349,
722
+ "eval_loss": 0.6003134250640869,
723
+ "eval_runtime": 968.0302,
724
+ "eval_samples_per_second": 6.329,
725
+ "eval_steps_per_second": 0.791,
726
+ "step": 2000
727
+ },
728
+ {
729
+ "epoch": 0.6532988357050453,
730
+ "grad_norm": 0.3548874258995056,
731
+ "learning_rate": 0.00014655172413793104,
732
+ "loss": 0.4548,
733
+ "step": 2020
734
+ },
735
+ {
736
+ "epoch": 0.6597671410090556,
737
+ "grad_norm": 0.3707274794578552,
738
+ "learning_rate": 0.00014584799437016187,
739
+ "loss": 0.4271,
740
+ "step": 2040
741
+ },
742
+ {
743
+ "epoch": 0.666235446313066,
744
+ "grad_norm": 0.37172695994377136,
745
+ "learning_rate": 0.0001451442646023927,
746
+ "loss": 0.4626,
747
+ "step": 2060
748
+ },
749
+ {
750
+ "epoch": 0.6727037516170763,
751
+ "grad_norm": 0.3459303081035614,
752
+ "learning_rate": 0.00014444053483462352,
753
+ "loss": 0.4317,
754
+ "step": 2080
755
+ },
756
+ {
757
+ "epoch": 0.6791720569210866,
758
+ "grad_norm": 0.3477213680744171,
759
+ "learning_rate": 0.00014373680506685435,
760
+ "loss": 0.4243,
761
+ "step": 2100
762
+ },
763
+ {
764
+ "epoch": 0.685640362225097,
765
+ "grad_norm": 0.31847599148750305,
766
+ "learning_rate": 0.00014303307529908517,
767
+ "loss": 0.4234,
768
+ "step": 2120
769
+ },
770
+ {
771
+ "epoch": 0.6921086675291074,
772
+ "grad_norm": 0.3231470584869385,
773
+ "learning_rate": 0.00014232934553131597,
774
+ "loss": 0.4302,
775
+ "step": 2140
776
+ },
777
+ {
778
+ "epoch": 0.6985769728331177,
779
+ "grad_norm": 0.3745759427547455,
780
+ "learning_rate": 0.0001416256157635468,
781
+ "loss": 0.4533,
782
+ "step": 2160
783
+ },
784
+ {
785
+ "epoch": 0.705045278137128,
786
+ "grad_norm": 0.34834805130958557,
787
+ "learning_rate": 0.00014092188599577763,
788
+ "loss": 0.4748,
789
+ "step": 2180
790
+ },
791
+ {
792
+ "epoch": 0.7115135834411385,
793
+ "grad_norm": 0.3675827085971832,
794
+ "learning_rate": 0.00014021815622800845,
795
+ "loss": 0.4401,
796
+ "step": 2200
797
+ },
798
+ {
799
+ "epoch": 0.7179818887451488,
800
+ "grad_norm": 0.2839801013469696,
801
+ "learning_rate": 0.00013951442646023928,
802
+ "loss": 0.4332,
803
+ "step": 2220
804
+ },
805
+ {
806
+ "epoch": 0.7244501940491591,
807
+ "grad_norm": 0.27037495374679565,
808
+ "learning_rate": 0.00013881069669247008,
809
+ "loss": 0.4382,
810
+ "step": 2240
811
+ },
812
+ {
813
+ "epoch": 0.7309184993531694,
814
+ "grad_norm": 0.34917208552360535,
815
+ "learning_rate": 0.00013810696692470093,
816
+ "loss": 0.4251,
817
+ "step": 2260
818
+ },
819
+ {
820
+ "epoch": 0.7373868046571799,
821
+ "grad_norm": 0.39801281690597534,
822
+ "learning_rate": 0.00013740323715693176,
823
+ "loss": 0.4432,
824
+ "step": 2280
825
+ },
826
+ {
827
+ "epoch": 0.7438551099611902,
828
+ "grad_norm": 0.8932256102561951,
829
+ "learning_rate": 0.00013669950738916259,
830
+ "loss": 0.4577,
831
+ "step": 2300
832
+ },
833
+ {
834
+ "epoch": 0.7503234152652005,
835
+ "grad_norm": 0.35125109553337097,
836
+ "learning_rate": 0.00013599577762139339,
837
+ "loss": 0.4502,
838
+ "step": 2320
839
+ },
840
+ {
841
+ "epoch": 0.7567917205692108,
842
+ "grad_norm": 0.3012046813964844,
843
+ "learning_rate": 0.0001352920478536242,
844
+ "loss": 0.4513,
845
+ "step": 2340
846
+ },
847
+ {
848
+ "epoch": 0.7632600258732212,
849
+ "grad_norm": 0.3242063820362091,
850
+ "learning_rate": 0.00013458831808585504,
851
+ "loss": 0.4559,
852
+ "step": 2360
853
+ },
854
+ {
855
+ "epoch": 0.7697283311772316,
856
+ "grad_norm": 0.3523860275745392,
857
+ "learning_rate": 0.00013388458831808587,
858
+ "loss": 0.4434,
859
+ "step": 2380
860
+ },
861
+ {
862
+ "epoch": 0.7761966364812419,
863
+ "grad_norm": 0.29160013794898987,
864
+ "learning_rate": 0.0001331808585503167,
865
+ "loss": 0.44,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 0.7826649417852523,
870
+ "grad_norm": 0.3047971725463867,
871
+ "learning_rate": 0.0001324771287825475,
872
+ "loss": 0.4336,
873
+ "step": 2420
874
+ },
875
+ {
876
+ "epoch": 0.7891332470892626,
877
+ "grad_norm": 0.3192126750946045,
878
+ "learning_rate": 0.00013177339901477832,
879
+ "loss": 0.4495,
880
+ "step": 2440
881
+ },
882
+ {
883
+ "epoch": 0.795601552393273,
884
+ "grad_norm": 0.37997299432754517,
885
+ "learning_rate": 0.00013106966924700917,
886
+ "loss": 0.4685,
887
+ "step": 2460
888
+ },
889
+ {
890
+ "epoch": 0.8020698576972833,
891
+ "grad_norm": 0.2811839282512665,
892
+ "learning_rate": 0.00013036593947924,
893
+ "loss": 0.4206,
894
+ "step": 2480
895
+ },
896
+ {
897
+ "epoch": 0.8085381630012937,
898
+ "grad_norm": 0.3121369779109955,
899
+ "learning_rate": 0.0001296622097114708,
900
+ "loss": 0.425,
901
+ "step": 2500
902
+ },
903
+ {
904
+ "epoch": 0.815006468305304,
905
+ "grad_norm": 0.3134746253490448,
906
+ "learning_rate": 0.00012895847994370162,
907
+ "loss": 0.4498,
908
+ "step": 2520
909
+ },
910
+ {
911
+ "epoch": 0.8214747736093143,
912
+ "grad_norm": 0.36752375960350037,
913
+ "learning_rate": 0.00012825475017593245,
914
+ "loss": 0.4288,
915
+ "step": 2540
916
+ },
917
+ {
918
+ "epoch": 0.8279430789133247,
919
+ "grad_norm": 0.3915141224861145,
920
+ "learning_rate": 0.00012755102040816328,
921
+ "loss": 0.4605,
922
+ "step": 2560
923
+ },
924
+ {
925
+ "epoch": 0.8344113842173351,
926
+ "grad_norm": 0.3662431538105011,
927
+ "learning_rate": 0.00012684729064039408,
928
+ "loss": 0.4194,
929
+ "step": 2580
930
+ },
931
+ {
932
+ "epoch": 0.8408796895213454,
933
+ "grad_norm": 0.3387574851512909,
934
+ "learning_rate": 0.0001261435608726249,
935
+ "loss": 0.4258,
936
+ "step": 2600
937
+ },
938
+ {
939
+ "epoch": 0.8473479948253557,
940
+ "grad_norm": 0.38025060296058655,
941
+ "learning_rate": 0.00012543983110485573,
942
+ "loss": 0.4291,
943
+ "step": 2620
944
+ },
945
+ {
946
+ "epoch": 0.8538163001293662,
947
+ "grad_norm": 0.3517441749572754,
948
+ "learning_rate": 0.00012473610133708658,
949
+ "loss": 0.4384,
950
+ "step": 2640
951
+ },
952
+ {
953
+ "epoch": 0.8602846054333765,
954
+ "grad_norm": 0.33037152886390686,
955
+ "learning_rate": 0.0001240323715693174,
956
+ "loss": 0.4452,
957
+ "step": 2660
958
+ },
959
+ {
960
+ "epoch": 0.8667529107373868,
961
+ "grad_norm": 0.3645654618740082,
962
+ "learning_rate": 0.0001233286418015482,
963
+ "loss": 0.3992,
964
+ "step": 2680
965
+ },
966
+ {
967
+ "epoch": 0.8732212160413971,
968
+ "grad_norm": 0.32396814227104187,
969
+ "learning_rate": 0.00012262491203377904,
970
+ "loss": 0.4,
971
+ "step": 2700
972
+ },
973
+ {
974
+ "epoch": 0.8796895213454075,
975
+ "grad_norm": 0.3377726078033447,
976
+ "learning_rate": 0.00012192118226600986,
977
+ "loss": 0.4306,
978
+ "step": 2720
979
+ },
980
+ {
981
+ "epoch": 0.8861578266494179,
982
+ "grad_norm": 0.32345983386039734,
983
+ "learning_rate": 0.00012121745249824067,
984
+ "loss": 0.4225,
985
+ "step": 2740
986
+ },
987
+ {
988
+ "epoch": 0.8926261319534282,
989
+ "grad_norm": 0.34778884053230286,
990
+ "learning_rate": 0.0001205137227304715,
991
+ "loss": 0.4256,
992
+ "step": 2760
993
+ },
994
+ {
995
+ "epoch": 0.8990944372574385,
996
+ "grad_norm": 0.3861066997051239,
997
+ "learning_rate": 0.00011980999296270233,
998
+ "loss": 0.4465,
999
+ "step": 2780
1000
+ },
1001
+ {
1002
+ "epoch": 0.9055627425614489,
1003
+ "grad_norm": 0.3882784843444824,
1004
+ "learning_rate": 0.00011910626319493314,
1005
+ "loss": 0.4026,
1006
+ "step": 2800
1007
+ },
1008
+ {
1009
+ "epoch": 0.9120310478654593,
1010
+ "grad_norm": 0.4380407929420471,
1011
+ "learning_rate": 0.00011840253342716397,
1012
+ "loss": 0.4375,
1013
+ "step": 2820
1014
+ },
1015
+ {
1016
+ "epoch": 0.9184993531694696,
1017
+ "grad_norm": 0.35943394899368286,
1018
+ "learning_rate": 0.00011769880365939481,
1019
+ "loss": 0.4337,
1020
+ "step": 2840
1021
+ },
1022
+ {
1023
+ "epoch": 0.92496765847348,
1024
+ "grad_norm": 0.3862978219985962,
1025
+ "learning_rate": 0.00011699507389162563,
1026
+ "loss": 0.3949,
1027
+ "step": 2860
1028
+ },
1029
+ {
1030
+ "epoch": 0.9314359637774903,
1031
+ "grad_norm": 0.34592369198799133,
1032
+ "learning_rate": 0.00011629134412385645,
1033
+ "loss": 0.3798,
1034
+ "step": 2880
1035
+ },
1036
+ {
1037
+ "epoch": 0.9379042690815006,
1038
+ "grad_norm": 0.34942877292633057,
1039
+ "learning_rate": 0.00011558761435608727,
1040
+ "loss": 0.4235,
1041
+ "step": 2900
1042
+ },
1043
+ {
1044
+ "epoch": 0.944372574385511,
1045
+ "grad_norm": 0.28018903732299805,
1046
+ "learning_rate": 0.00011488388458831809,
1047
+ "loss": 0.4353,
1048
+ "step": 2920
1049
+ },
1050
+ {
1051
+ "epoch": 0.9508408796895214,
1052
+ "grad_norm": 0.3505672812461853,
1053
+ "learning_rate": 0.00011418015482054891,
1054
+ "loss": 0.3954,
1055
+ "step": 2940
1056
+ },
1057
+ {
1058
+ "epoch": 0.9573091849935317,
1059
+ "grad_norm": 0.3264220654964447,
1060
+ "learning_rate": 0.00011347642505277974,
1061
+ "loss": 0.4422,
1062
+ "step": 2960
1063
+ },
1064
+ {
1065
+ "epoch": 0.963777490297542,
1066
+ "grad_norm": 0.36382535099983215,
1067
+ "learning_rate": 0.00011277269528501055,
1068
+ "loss": 0.4116,
1069
+ "step": 2980
1070
+ },
1071
+ {
1072
+ "epoch": 0.9702457956015524,
1073
+ "grad_norm": 0.3051494359970093,
1074
+ "learning_rate": 0.00011206896551724138,
1075
+ "loss": 0.4135,
1076
+ "step": 3000
1077
+ },
1078
+ {
1079
+ "epoch": 0.9702457956015524,
1080
+ "eval_loss": 0.5850913524627686,
1081
+ "eval_runtime": 967.6256,
1082
+ "eval_samples_per_second": 6.332,
1083
+ "eval_steps_per_second": 0.792,
1084
+ "step": 3000
1085
+ },
1086
+ {
1087
+ "epoch": 0.9767141009055628,
1088
+ "grad_norm": 0.34533271193504333,
1089
+ "learning_rate": 0.00011136523574947222,
1090
+ "loss": 0.4312,
1091
+ "step": 3020
1092
+ },
1093
+ {
1094
+ "epoch": 0.9831824062095731,
1095
+ "grad_norm": 0.35221609473228455,
1096
+ "learning_rate": 0.00011066150598170305,
1097
+ "loss": 0.4245,
1098
+ "step": 3040
1099
+ },
1100
+ {
1101
+ "epoch": 0.9896507115135834,
1102
+ "grad_norm": 0.3893713057041168,
1103
+ "learning_rate": 0.00010995777621393386,
1104
+ "loss": 0.4048,
1105
+ "step": 3060
1106
+ },
1107
+ {
1108
+ "epoch": 0.9961190168175937,
1109
+ "grad_norm": 0.3586278557777405,
1110
+ "learning_rate": 0.00010925404644616468,
1111
+ "loss": 0.4159,
1112
+ "step": 3080
1113
+ },
1114
+ {
1115
+ "epoch": 1.0025873221216042,
1116
+ "grad_norm": 0.32620444893836975,
1117
+ "learning_rate": 0.0001085503166783955,
1118
+ "loss": 0.3741,
1119
+ "step": 3100
1120
+ },
1121
+ {
1122
+ "epoch": 1.0090556274256144,
1123
+ "grad_norm": 0.33812645077705383,
1124
+ "learning_rate": 0.00010784658691062632,
1125
+ "loss": 0.3694,
1126
+ "step": 3120
1127
+ },
1128
+ {
1129
+ "epoch": 1.0155239327296248,
1130
+ "grad_norm": 0.39038708806037903,
1131
+ "learning_rate": 0.00010714285714285715,
1132
+ "loss": 0.3882,
1133
+ "step": 3140
1134
+ },
1135
+ {
1136
+ "epoch": 1.0219922380336353,
1137
+ "grad_norm": 0.3209969997406006,
1138
+ "learning_rate": 0.00010643912737508796,
1139
+ "loss": 0.3437,
1140
+ "step": 3160
1141
+ },
1142
+ {
1143
+ "epoch": 1.0284605433376455,
1144
+ "grad_norm": 0.3705078661441803,
1145
+ "learning_rate": 0.00010573539760731879,
1146
+ "loss": 0.3568,
1147
+ "step": 3180
1148
+ },
1149
+ {
1150
+ "epoch": 1.034928848641656,
1151
+ "grad_norm": 0.3562203645706177,
1152
+ "learning_rate": 0.0001050316678395496,
1153
+ "loss": 0.3916,
1154
+ "step": 3200
1155
+ },
1156
+ {
1157
+ "epoch": 1.0413971539456663,
1158
+ "grad_norm": 0.3724074363708496,
1159
+ "learning_rate": 0.00010432793807178046,
1160
+ "loss": 0.3866,
1161
+ "step": 3220
1162
+ },
1163
+ {
1164
+ "epoch": 1.0478654592496766,
1165
+ "grad_norm": 0.40453559160232544,
1166
+ "learning_rate": 0.00010362420830401127,
1167
+ "loss": 0.368,
1168
+ "step": 3240
1169
+ },
1170
+ {
1171
+ "epoch": 1.054333764553687,
1172
+ "grad_norm": 0.3316144049167633,
1173
+ "learning_rate": 0.0001029204785362421,
1174
+ "loss": 0.3696,
1175
+ "step": 3260
1176
+ },
1177
+ {
1178
+ "epoch": 1.0608020698576972,
1179
+ "grad_norm": 0.3503650724887848,
1180
+ "learning_rate": 0.00010221674876847291,
1181
+ "loss": 0.3644,
1182
+ "step": 3280
1183
+ },
1184
+ {
1185
+ "epoch": 1.0672703751617076,
1186
+ "grad_norm": 0.3136923909187317,
1187
+ "learning_rate": 0.00010151301900070374,
1188
+ "loss": 0.3414,
1189
+ "step": 3300
1190
+ },
1191
+ {
1192
+ "epoch": 1.073738680465718,
1193
+ "grad_norm": 0.3573864996433258,
1194
+ "learning_rate": 0.00010080928923293456,
1195
+ "loss": 0.3586,
1196
+ "step": 3320
1197
+ },
1198
+ {
1199
+ "epoch": 1.0802069857697283,
1200
+ "grad_norm": 0.26215964555740356,
1201
+ "learning_rate": 0.00010010555946516538,
1202
+ "loss": 0.356,
1203
+ "step": 3340
1204
+ },
1205
+ {
1206
+ "epoch": 1.0866752910737387,
1207
+ "grad_norm": 0.3103466331958771,
1208
+ "learning_rate": 9.940182969739622e-05,
1209
+ "loss": 0.3478,
1210
+ "step": 3360
1211
+ },
1212
+ {
1213
+ "epoch": 1.093143596377749,
1214
+ "grad_norm": 0.3601842224597931,
1215
+ "learning_rate": 9.869809992962703e-05,
1216
+ "loss": 0.3635,
1217
+ "step": 3380
1218
+ },
1219
+ {
1220
+ "epoch": 1.0996119016817594,
1221
+ "grad_norm": 0.3715918958187103,
1222
+ "learning_rate": 9.799437016185785e-05,
1223
+ "loss": 0.3489,
1224
+ "step": 3400
1225
+ },
1226
+ {
1227
+ "epoch": 1.1060802069857698,
1228
+ "grad_norm": 0.4216213524341583,
1229
+ "learning_rate": 9.729064039408867e-05,
1230
+ "loss": 0.3461,
1231
+ "step": 3420
1232
+ },
1233
+ {
1234
+ "epoch": 1.11254851228978,
1235
+ "grad_norm": 0.3878704905509949,
1236
+ "learning_rate": 9.65869106263195e-05,
1237
+ "loss": 0.3548,
1238
+ "step": 3440
1239
+ },
1240
+ {
1241
+ "epoch": 1.1190168175937905,
1242
+ "grad_norm": 0.36658766865730286,
1243
+ "learning_rate": 9.588318085855032e-05,
1244
+ "loss": 0.3395,
1245
+ "step": 3460
1246
+ },
1247
+ {
1248
+ "epoch": 1.1254851228978007,
1249
+ "grad_norm": 0.38134855031967163,
1250
+ "learning_rate": 9.517945109078115e-05,
1251
+ "loss": 0.3829,
1252
+ "step": 3480
1253
+ },
1254
+ {
1255
+ "epoch": 1.131953428201811,
1256
+ "grad_norm": 0.42683905363082886,
1257
+ "learning_rate": 9.447572132301197e-05,
1258
+ "loss": 0.3914,
1259
+ "step": 3500
1260
+ },
1261
+ {
1262
+ "epoch": 1.1384217335058215,
1263
+ "grad_norm": 0.3340086042881012,
1264
+ "learning_rate": 9.377199155524279e-05,
1265
+ "loss": 0.3301,
1266
+ "step": 3520
1267
+ },
1268
+ {
1269
+ "epoch": 1.1448900388098318,
1270
+ "grad_norm": 0.37897104024887085,
1271
+ "learning_rate": 9.306826178747361e-05,
1272
+ "loss": 0.3536,
1273
+ "step": 3540
1274
+ },
1275
+ {
1276
+ "epoch": 1.1513583441138422,
1277
+ "grad_norm": 0.361061692237854,
1278
+ "learning_rate": 9.236453201970444e-05,
1279
+ "loss": 0.338,
1280
+ "step": 3560
1281
+ },
1282
+ {
1283
+ "epoch": 1.1578266494178524,
1284
+ "grad_norm": 0.34730997681617737,
1285
+ "learning_rate": 9.166080225193527e-05,
1286
+ "loss": 0.3679,
1287
+ "step": 3580
1288
+ },
1289
+ {
1290
+ "epoch": 1.1642949547218628,
1291
+ "grad_norm": 0.29455122351646423,
1292
+ "learning_rate": 9.095707248416608e-05,
1293
+ "loss": 0.3602,
1294
+ "step": 3600
1295
+ },
1296
+ {
1297
+ "epoch": 1.1707632600258733,
1298
+ "grad_norm": 0.33159056305885315,
1299
+ "learning_rate": 9.02533427163969e-05,
1300
+ "loss": 0.3359,
1301
+ "step": 3620
1302
+ },
1303
+ {
1304
+ "epoch": 1.1772315653298835,
1305
+ "grad_norm": 0.48726096749305725,
1306
+ "learning_rate": 8.954961294862773e-05,
1307
+ "loss": 0.3879,
1308
+ "step": 3640
1309
+ },
1310
+ {
1311
+ "epoch": 1.183699870633894,
1312
+ "grad_norm": 0.3182481825351715,
1313
+ "learning_rate": 8.884588318085856e-05,
1314
+ "loss": 0.3392,
1315
+ "step": 3660
1316
+ },
1317
+ {
1318
+ "epoch": 1.1901681759379044,
1319
+ "grad_norm": 0.3488766551017761,
1320
+ "learning_rate": 8.814215341308939e-05,
1321
+ "loss": 0.3419,
1322
+ "step": 3680
1323
+ },
1324
+ {
1325
+ "epoch": 1.1966364812419146,
1326
+ "grad_norm": 0.34345540404319763,
1327
+ "learning_rate": 8.74384236453202e-05,
1328
+ "loss": 0.3598,
1329
+ "step": 3700
1330
+ },
1331
+ {
1332
+ "epoch": 1.203104786545925,
1333
+ "grad_norm": 0.37642374634742737,
1334
+ "learning_rate": 8.673469387755102e-05,
1335
+ "loss": 0.3578,
1336
+ "step": 3720
1337
+ },
1338
+ {
1339
+ "epoch": 1.2095730918499352,
1340
+ "grad_norm": 0.3726005554199219,
1341
+ "learning_rate": 8.603096410978185e-05,
1342
+ "loss": 0.3547,
1343
+ "step": 3740
1344
+ },
1345
+ {
1346
+ "epoch": 1.2160413971539457,
1347
+ "grad_norm": 0.36603832244873047,
1348
+ "learning_rate": 8.532723434201268e-05,
1349
+ "loss": 0.3831,
1350
+ "step": 3760
1351
+ },
1352
+ {
1353
+ "epoch": 1.222509702457956,
1354
+ "grad_norm": 0.345103919506073,
1355
+ "learning_rate": 8.462350457424349e-05,
1356
+ "loss": 0.3563,
1357
+ "step": 3780
1358
+ },
1359
+ {
1360
+ "epoch": 1.2289780077619663,
1361
+ "grad_norm": 0.31071504950523376,
1362
+ "learning_rate": 8.391977480647432e-05,
1363
+ "loss": 0.3263,
1364
+ "step": 3800
1365
+ },
1366
+ {
1367
+ "epoch": 1.2354463130659767,
1368
+ "grad_norm": 0.3534263074398041,
1369
+ "learning_rate": 8.321604503870513e-05,
1370
+ "loss": 0.3591,
1371
+ "step": 3820
1372
+ },
1373
+ {
1374
+ "epoch": 1.2419146183699872,
1375
+ "grad_norm": 0.3568720817565918,
1376
+ "learning_rate": 8.251231527093597e-05,
1377
+ "loss": 0.3815,
1378
+ "step": 3840
1379
+ },
1380
+ {
1381
+ "epoch": 1.2483829236739974,
1382
+ "grad_norm": 0.3450963795185089,
1383
+ "learning_rate": 8.18085855031668e-05,
1384
+ "loss": 0.3481,
1385
+ "step": 3860
1386
+ },
1387
+ {
1388
+ "epoch": 1.2548512289780078,
1389
+ "grad_norm": 0.3554648160934448,
1390
+ "learning_rate": 8.110485573539761e-05,
1391
+ "loss": 0.3364,
1392
+ "step": 3880
1393
+ },
1394
+ {
1395
+ "epoch": 1.261319534282018,
1396
+ "grad_norm": 0.4188801348209381,
1397
+ "learning_rate": 8.040112596762844e-05,
1398
+ "loss": 0.3686,
1399
+ "step": 3900
1400
+ },
1401
+ {
1402
+ "epoch": 1.2677878395860285,
1403
+ "grad_norm": 0.3129253089427948,
1404
+ "learning_rate": 7.969739619985925e-05,
1405
+ "loss": 0.3664,
1406
+ "step": 3920
1407
+ },
1408
+ {
1409
+ "epoch": 1.274256144890039,
1410
+ "grad_norm": 0.3885388970375061,
1411
+ "learning_rate": 7.899366643209009e-05,
1412
+ "loss": 0.3415,
1413
+ "step": 3940
1414
+ },
1415
+ {
1416
+ "epoch": 1.2807244501940491,
1417
+ "grad_norm": 0.4131453335285187,
1418
+ "learning_rate": 7.82899366643209e-05,
1419
+ "loss": 0.3552,
1420
+ "step": 3960
1421
+ },
1422
+ {
1423
+ "epoch": 1.2871927554980596,
1424
+ "grad_norm": 0.39486396312713623,
1425
+ "learning_rate": 7.758620689655173e-05,
1426
+ "loss": 0.366,
1427
+ "step": 3980
1428
+ },
1429
+ {
1430
+ "epoch": 1.2936610608020698,
1431
+ "grad_norm": 0.3580794930458069,
1432
+ "learning_rate": 7.688247712878254e-05,
1433
+ "loss": 0.3585,
1434
+ "step": 4000
1435
+ },
1436
+ {
1437
+ "epoch": 1.2936610608020698,
1438
+ "eval_loss": 0.5952740907669067,
1439
+ "eval_runtime": 967.8131,
1440
+ "eval_samples_per_second": 6.331,
1441
+ "eval_steps_per_second": 0.791,
1442
+ "step": 4000
1443
+ },
1444
+ {
1445
+ "epoch": 1.3001293661060802,
1446
+ "grad_norm": 0.3903474509716034,
1447
+ "learning_rate": 7.617874736101337e-05,
1448
+ "loss": 0.3347,
1449
+ "step": 4020
1450
+ },
1451
+ {
1452
+ "epoch": 1.3065976714100906,
1453
+ "grad_norm": 0.4154898226261139,
1454
+ "learning_rate": 7.54750175932442e-05,
1455
+ "loss": 0.3314,
1456
+ "step": 4040
1457
+ },
1458
+ {
1459
+ "epoch": 1.3130659767141009,
1460
+ "grad_norm": 0.33763521909713745,
1461
+ "learning_rate": 7.477128782547502e-05,
1462
+ "loss": 0.3552,
1463
+ "step": 4060
1464
+ },
1465
+ {
1466
+ "epoch": 1.3195342820181113,
1467
+ "grad_norm": 0.5618635416030884,
1468
+ "learning_rate": 7.406755805770585e-05,
1469
+ "loss": 0.363,
1470
+ "step": 4080
1471
+ },
1472
+ {
1473
+ "epoch": 1.3260025873221215,
1474
+ "grad_norm": 0.37114137411117554,
1475
+ "learning_rate": 7.336382828993666e-05,
1476
+ "loss": 0.3421,
1477
+ "step": 4100
1478
+ },
1479
+ {
1480
+ "epoch": 1.332470892626132,
1481
+ "grad_norm": 0.34323641657829285,
1482
+ "learning_rate": 7.266009852216749e-05,
1483
+ "loss": 0.3659,
1484
+ "step": 4120
1485
+ },
1486
+ {
1487
+ "epoch": 1.3389391979301424,
1488
+ "grad_norm": 0.3527199625968933,
1489
+ "learning_rate": 7.195636875439831e-05,
1490
+ "loss": 0.3412,
1491
+ "step": 4140
1492
+ },
1493
+ {
1494
+ "epoch": 1.3454075032341526,
1495
+ "grad_norm": 0.3413406014442444,
1496
+ "learning_rate": 7.125263898662914e-05,
1497
+ "loss": 0.338,
1498
+ "step": 4160
1499
+ },
1500
+ {
1501
+ "epoch": 1.351875808538163,
1502
+ "grad_norm": 0.34787651896476746,
1503
+ "learning_rate": 7.054890921885995e-05,
1504
+ "loss": 0.3681,
1505
+ "step": 4180
1506
+ },
1507
+ {
1508
+ "epoch": 1.3583441138421732,
1509
+ "grad_norm": 0.2709845006465912,
1510
+ "learning_rate": 6.984517945109078e-05,
1511
+ "loss": 0.325,
1512
+ "step": 4200
1513
+ },
1514
+ {
1515
+ "epoch": 1.3648124191461837,
1516
+ "grad_norm": 0.30966514348983765,
1517
+ "learning_rate": 6.91414496833216e-05,
1518
+ "loss": 0.3677,
1519
+ "step": 4220
1520
+ },
1521
+ {
1522
+ "epoch": 1.371280724450194,
1523
+ "grad_norm": 0.35980409383773804,
1524
+ "learning_rate": 6.843771991555243e-05,
1525
+ "loss": 0.335,
1526
+ "step": 4240
1527
+ },
1528
+ {
1529
+ "epoch": 1.3777490297542043,
1530
+ "grad_norm": 0.5173650979995728,
1531
+ "learning_rate": 6.773399014778326e-05,
1532
+ "loss": 0.348,
1533
+ "step": 4260
1534
+ },
1535
+ {
1536
+ "epoch": 1.3842173350582148,
1537
+ "grad_norm": 0.3048713207244873,
1538
+ "learning_rate": 6.703026038001407e-05,
1539
+ "loss": 0.3323,
1540
+ "step": 4280
1541
+ },
1542
+ {
1543
+ "epoch": 1.390685640362225,
1544
+ "grad_norm": 0.3427985906600952,
1545
+ "learning_rate": 6.63265306122449e-05,
1546
+ "loss": 0.3571,
1547
+ "step": 4300
1548
+ },
1549
+ {
1550
+ "epoch": 1.3971539456662354,
1551
+ "grad_norm": 0.406012624502182,
1552
+ "learning_rate": 6.562280084447573e-05,
1553
+ "loss": 0.3546,
1554
+ "step": 4320
1555
+ },
1556
+ {
1557
+ "epoch": 1.4036222509702458,
1558
+ "grad_norm": 0.31981325149536133,
1559
+ "learning_rate": 6.491907107670655e-05,
1560
+ "loss": 0.3281,
1561
+ "step": 4340
1562
+ },
1563
+ {
1564
+ "epoch": 1.4100905562742563,
1565
+ "grad_norm": 0.3936704099178314,
1566
+ "learning_rate": 6.421534130893737e-05,
1567
+ "loss": 0.3444,
1568
+ "step": 4360
1569
+ },
1570
+ {
1571
+ "epoch": 1.4165588615782665,
1572
+ "grad_norm": 0.3658936023712158,
1573
+ "learning_rate": 6.351161154116819e-05,
1574
+ "loss": 0.339,
1575
+ "step": 4380
1576
+ },
1577
+ {
1578
+ "epoch": 1.4230271668822767,
1579
+ "grad_norm": 0.32523271441459656,
1580
+ "learning_rate": 6.280788177339902e-05,
1581
+ "loss": 0.3311,
1582
+ "step": 4400
1583
+ },
1584
+ {
1585
+ "epoch": 1.4294954721862871,
1586
+ "grad_norm": 0.4120417833328247,
1587
+ "learning_rate": 6.210415200562984e-05,
1588
+ "loss": 0.3342,
1589
+ "step": 4420
1590
+ },
1591
+ {
1592
+ "epoch": 1.4359637774902976,
1593
+ "grad_norm": 0.35293978452682495,
1594
+ "learning_rate": 6.140042223786067e-05,
1595
+ "loss": 0.3561,
1596
+ "step": 4440
1597
+ },
1598
+ {
1599
+ "epoch": 1.442432082794308,
1600
+ "grad_norm": 0.33231040835380554,
1601
+ "learning_rate": 6.0696692470091484e-05,
1602
+ "loss": 0.3573,
1603
+ "step": 4460
1604
+ },
1605
+ {
1606
+ "epoch": 1.4489003880983182,
1607
+ "grad_norm": 0.4345926344394684,
1608
+ "learning_rate": 5.999296270232231e-05,
1609
+ "loss": 0.3364,
1610
+ "step": 4480
1611
+ },
1612
+ {
1613
+ "epoch": 1.4553686934023287,
1614
+ "grad_norm": 0.34958288073539734,
1615
+ "learning_rate": 5.928923293455313e-05,
1616
+ "loss": 0.3162,
1617
+ "step": 4500
1618
+ },
1619
+ {
1620
+ "epoch": 1.4618369987063389,
1621
+ "grad_norm": 0.4669477045536041,
1622
+ "learning_rate": 5.8585503166783964e-05,
1623
+ "loss": 0.3455,
1624
+ "step": 4520
1625
+ },
1626
+ {
1627
+ "epoch": 1.4683053040103493,
1628
+ "grad_norm": 0.32660993933677673,
1629
+ "learning_rate": 5.788177339901478e-05,
1630
+ "loss": 0.3228,
1631
+ "step": 4540
1632
+ },
1633
+ {
1634
+ "epoch": 1.4747736093143597,
1635
+ "grad_norm": 0.3654051423072815,
1636
+ "learning_rate": 5.71780436312456e-05,
1637
+ "loss": 0.3293,
1638
+ "step": 4560
1639
+ },
1640
+ {
1641
+ "epoch": 1.48124191461837,
1642
+ "grad_norm": 0.3496916592121124,
1643
+ "learning_rate": 5.647431386347642e-05,
1644
+ "loss": 0.3382,
1645
+ "step": 4580
1646
+ },
1647
+ {
1648
+ "epoch": 1.4877102199223804,
1649
+ "grad_norm": 0.3890388309955597,
1650
+ "learning_rate": 5.577058409570724e-05,
1651
+ "loss": 0.3468,
1652
+ "step": 4600
1653
+ },
1654
+ {
1655
+ "epoch": 1.4941785252263906,
1656
+ "grad_norm": 0.3924092650413513,
1657
+ "learning_rate": 5.5066854327938076e-05,
1658
+ "loss": 0.3467,
1659
+ "step": 4620
1660
+ },
1661
+ {
1662
+ "epoch": 1.500646830530401,
1663
+ "grad_norm": 0.4015033543109894,
1664
+ "learning_rate": 5.4363124560168896e-05,
1665
+ "loss": 0.3478,
1666
+ "step": 4640
1667
+ },
1668
+ {
1669
+ "epoch": 1.5071151358344115,
1670
+ "grad_norm": 0.34742093086242676,
1671
+ "learning_rate": 5.365939479239972e-05,
1672
+ "loss": 0.3252,
1673
+ "step": 4660
1674
+ },
1675
+ {
1676
+ "epoch": 1.5135834411384217,
1677
+ "grad_norm": 0.3477989435195923,
1678
+ "learning_rate": 5.295566502463054e-05,
1679
+ "loss": 0.3627,
1680
+ "step": 4680
1681
+ },
1682
+ {
1683
+ "epoch": 1.5200517464424321,
1684
+ "grad_norm": 0.4194832146167755,
1685
+ "learning_rate": 5.2251935256861375e-05,
1686
+ "loss": 0.3403,
1687
+ "step": 4700
1688
+ },
1689
+ {
1690
+ "epoch": 1.5265200517464423,
1691
+ "grad_norm": 0.37723594903945923,
1692
+ "learning_rate": 5.1548205489092195e-05,
1693
+ "loss": 0.3424,
1694
+ "step": 4720
1695
+ },
1696
+ {
1697
+ "epoch": 1.5329883570504528,
1698
+ "grad_norm": 0.3695043921470642,
1699
+ "learning_rate": 5.0844475721323015e-05,
1700
+ "loss": 0.3339,
1701
+ "step": 4740
1702
+ },
1703
+ {
1704
+ "epoch": 1.5394566623544632,
1705
+ "grad_norm": 0.35015031695365906,
1706
+ "learning_rate": 5.0140745953553834e-05,
1707
+ "loss": 0.3262,
1708
+ "step": 4760
1709
+ },
1710
+ {
1711
+ "epoch": 1.5459249676584734,
1712
+ "grad_norm": 0.39946597814559937,
1713
+ "learning_rate": 4.943701618578466e-05,
1714
+ "loss": 0.3649,
1715
+ "step": 4780
1716
+ },
1717
+ {
1718
+ "epoch": 1.5523932729624839,
1719
+ "grad_norm": 0.3246670961380005,
1720
+ "learning_rate": 4.873328641801548e-05,
1721
+ "loss": 0.3241,
1722
+ "step": 4800
1723
+ },
1724
+ {
1725
+ "epoch": 1.558861578266494,
1726
+ "grad_norm": 0.41913464665412903,
1727
+ "learning_rate": 4.802955665024631e-05,
1728
+ "loss": 0.3486,
1729
+ "step": 4820
1730
+ },
1731
+ {
1732
+ "epoch": 1.5653298835705045,
1733
+ "grad_norm": 0.38162580132484436,
1734
+ "learning_rate": 4.7325826882477134e-05,
1735
+ "loss": 0.3571,
1736
+ "step": 4840
1737
+ },
1738
+ {
1739
+ "epoch": 1.571798188874515,
1740
+ "grad_norm": 0.34426695108413696,
1741
+ "learning_rate": 4.6622097114707953e-05,
1742
+ "loss": 0.3232,
1743
+ "step": 4860
1744
+ },
1745
+ {
1746
+ "epoch": 1.5782664941785254,
1747
+ "grad_norm": 0.3782813847064972,
1748
+ "learning_rate": 4.591836734693878e-05,
1749
+ "loss": 0.346,
1750
+ "step": 4880
1751
+ },
1752
+ {
1753
+ "epoch": 1.5847347994825356,
1754
+ "grad_norm": 0.42736709117889404,
1755
+ "learning_rate": 4.52146375791696e-05,
1756
+ "loss": 0.3275,
1757
+ "step": 4900
1758
+ },
1759
+ {
1760
+ "epoch": 1.5912031047865458,
1761
+ "grad_norm": 0.37077978253364563,
1762
+ "learning_rate": 4.4510907811400426e-05,
1763
+ "loss": 0.3312,
1764
+ "step": 4920
1765
+ },
1766
+ {
1767
+ "epoch": 1.5976714100905562,
1768
+ "grad_norm": 0.3196958303451538,
1769
+ "learning_rate": 4.3807178043631246e-05,
1770
+ "loss": 0.3156,
1771
+ "step": 4940
1772
+ },
1773
+ {
1774
+ "epoch": 1.6041397153945667,
1775
+ "grad_norm": 0.36680254340171814,
1776
+ "learning_rate": 4.3103448275862066e-05,
1777
+ "loss": 0.3293,
1778
+ "step": 4960
1779
+ },
1780
+ {
1781
+ "epoch": 1.6106080206985771,
1782
+ "grad_norm": 0.44455257058143616,
1783
+ "learning_rate": 4.239971850809289e-05,
1784
+ "loss": 0.3512,
1785
+ "step": 4980
1786
+ },
1787
+ {
1788
+ "epoch": 1.6170763260025873,
1789
+ "grad_norm": 0.37418290972709656,
1790
+ "learning_rate": 4.169598874032372e-05,
1791
+ "loss": 0.3422,
1792
+ "step": 5000
1793
+ },
1794
+ {
1795
+ "epoch": 1.6170763260025873,
1796
+ "eval_loss": 0.5881822109222412,
1797
+ "eval_runtime": 967.8912,
1798
+ "eval_samples_per_second": 6.33,
1799
+ "eval_steps_per_second": 0.791,
1800
+ "step": 5000
1801
+ },
1802
+ {
1803
+ "epoch": 1.6235446313065975,
1804
+ "grad_norm": 0.3209955096244812,
1805
+ "learning_rate": 4.0992258972554545e-05,
1806
+ "loss": 0.3158,
1807
+ "step": 5020
1808
+ },
1809
+ {
1810
+ "epoch": 1.630012936610608,
1811
+ "grad_norm": 0.31715771555900574,
1812
+ "learning_rate": 4.0288529204785365e-05,
1813
+ "loss": 0.3167,
1814
+ "step": 5040
1815
+ },
1816
+ {
1817
+ "epoch": 1.6364812419146184,
1818
+ "grad_norm": 0.3968450725078583,
1819
+ "learning_rate": 3.958479943701619e-05,
1820
+ "loss": 0.3219,
1821
+ "step": 5060
1822
+ },
1823
+ {
1824
+ "epoch": 1.6429495472186288,
1825
+ "grad_norm": 0.3187231123447418,
1826
+ "learning_rate": 3.888106966924701e-05,
1827
+ "loss": 0.3288,
1828
+ "step": 5080
1829
+ },
1830
+ {
1831
+ "epoch": 1.649417852522639,
1832
+ "grad_norm": 0.35491153597831726,
1833
+ "learning_rate": 3.817733990147783e-05,
1834
+ "loss": 0.3298,
1835
+ "step": 5100
1836
+ },
1837
+ {
1838
+ "epoch": 1.6558861578266493,
1839
+ "grad_norm": 0.4044147729873657,
1840
+ "learning_rate": 3.747361013370866e-05,
1841
+ "loss": 0.3505,
1842
+ "step": 5120
1843
+ },
1844
+ {
1845
+ "epoch": 1.6623544631306597,
1846
+ "grad_norm": 0.33881333470344543,
1847
+ "learning_rate": 3.676988036593948e-05,
1848
+ "loss": 0.3269,
1849
+ "step": 5140
1850
+ },
1851
+ {
1852
+ "epoch": 1.6688227684346701,
1853
+ "grad_norm": 0.3626604378223419,
1854
+ "learning_rate": 3.6066150598170304e-05,
1855
+ "loss": 0.3319,
1856
+ "step": 5160
1857
+ },
1858
+ {
1859
+ "epoch": 1.6752910737386806,
1860
+ "grad_norm": 0.39335256814956665,
1861
+ "learning_rate": 3.536242083040113e-05,
1862
+ "loss": 0.3234,
1863
+ "step": 5180
1864
+ },
1865
+ {
1866
+ "epoch": 1.6817593790426908,
1867
+ "grad_norm": 0.35490670800209045,
1868
+ "learning_rate": 3.465869106263195e-05,
1869
+ "loss": 0.3357,
1870
+ "step": 5200
1871
+ },
1872
+ {
1873
+ "epoch": 1.688227684346701,
1874
+ "grad_norm": 0.303774356842041,
1875
+ "learning_rate": 3.3954961294862776e-05,
1876
+ "loss": 0.3609,
1877
+ "step": 5220
1878
+ },
1879
+ {
1880
+ "epoch": 1.6946959896507114,
1881
+ "grad_norm": 0.3438192307949066,
1882
+ "learning_rate": 3.3251231527093596e-05,
1883
+ "loss": 0.3351,
1884
+ "step": 5240
1885
+ },
1886
+ {
1887
+ "epoch": 1.7011642949547219,
1888
+ "grad_norm": 0.3261934518814087,
1889
+ "learning_rate": 3.254750175932442e-05,
1890
+ "loss": 0.3192,
1891
+ "step": 5260
1892
+ },
1893
+ {
1894
+ "epoch": 1.7076326002587323,
1895
+ "grad_norm": 0.39429745078086853,
1896
+ "learning_rate": 3.184377199155524e-05,
1897
+ "loss": 0.3479,
1898
+ "step": 5280
1899
+ },
1900
+ {
1901
+ "epoch": 1.7141009055627425,
1902
+ "grad_norm": 0.4267764985561371,
1903
+ "learning_rate": 3.114004222378607e-05,
1904
+ "loss": 0.3306,
1905
+ "step": 5300
1906
+ },
1907
+ {
1908
+ "epoch": 1.720569210866753,
1909
+ "grad_norm": 0.3425467908382416,
1910
+ "learning_rate": 3.0436312456016892e-05,
1911
+ "loss": 0.3356,
1912
+ "step": 5320
1913
+ },
1914
+ {
1915
+ "epoch": 1.7270375161707632,
1916
+ "grad_norm": 0.3064535856246948,
1917
+ "learning_rate": 2.9732582688247712e-05,
1918
+ "loss": 0.3341,
1919
+ "step": 5340
1920
+ },
1921
+ {
1922
+ "epoch": 1.7335058214747736,
1923
+ "grad_norm": 0.3956790566444397,
1924
+ "learning_rate": 2.902885292047854e-05,
1925
+ "loss": 0.3375,
1926
+ "step": 5360
1927
+ },
1928
+ {
1929
+ "epoch": 1.739974126778784,
1930
+ "grad_norm": 0.36128687858581543,
1931
+ "learning_rate": 2.8325123152709358e-05,
1932
+ "loss": 0.3395,
1933
+ "step": 5380
1934
+ },
1935
+ {
1936
+ "epoch": 1.7464424320827943,
1937
+ "grad_norm": 0.33011215925216675,
1938
+ "learning_rate": 2.7621393384940185e-05,
1939
+ "loss": 0.3078,
1940
+ "step": 5400
1941
+ },
1942
+ {
1943
+ "epoch": 1.7529107373868047,
1944
+ "grad_norm": 0.3741130530834198,
1945
+ "learning_rate": 2.6917663617171008e-05,
1946
+ "loss": 0.362,
1947
+ "step": 5420
1948
+ },
1949
+ {
1950
+ "epoch": 1.759379042690815,
1951
+ "grad_norm": 0.319671094417572,
1952
+ "learning_rate": 2.6213933849401827e-05,
1953
+ "loss": 0.3105,
1954
+ "step": 5440
1955
+ },
1956
+ {
1957
+ "epoch": 1.7658473479948253,
1958
+ "grad_norm": 0.4425293505191803,
1959
+ "learning_rate": 2.5510204081632654e-05,
1960
+ "loss": 0.3327,
1961
+ "step": 5460
1962
+ },
1963
+ {
1964
+ "epoch": 1.7723156532988358,
1965
+ "grad_norm": 0.43660980463027954,
1966
+ "learning_rate": 2.4806474313863477e-05,
1967
+ "loss": 0.3057,
1968
+ "step": 5480
1969
+ },
1970
+ {
1971
+ "epoch": 1.7787839586028462,
1972
+ "grad_norm": 0.3476182222366333,
1973
+ "learning_rate": 2.41027445460943e-05,
1974
+ "loss": 0.3362,
1975
+ "step": 5500
1976
+ },
1977
+ {
1978
+ "epoch": 1.7852522639068564,
1979
+ "grad_norm": 0.37895113229751587,
1980
+ "learning_rate": 2.3399014778325123e-05,
1981
+ "loss": 0.337,
1982
+ "step": 5520
1983
+ },
1984
+ {
1985
+ "epoch": 1.7917205692108666,
1986
+ "grad_norm": 0.38524091243743896,
1987
+ "learning_rate": 2.2695285010555947e-05,
1988
+ "loss": 0.362,
1989
+ "step": 5540
1990
+ },
1991
+ {
1992
+ "epoch": 1.798188874514877,
1993
+ "grad_norm": 0.38090500235557556,
1994
+ "learning_rate": 2.199155524278677e-05,
1995
+ "loss": 0.3268,
1996
+ "step": 5560
1997
+ },
1998
+ {
1999
+ "epoch": 1.8046571798188875,
2000
+ "grad_norm": 0.4094742238521576,
2001
+ "learning_rate": 2.1287825475017596e-05,
2002
+ "loss": 0.3353,
2003
+ "step": 5580
2004
+ },
2005
+ {
2006
+ "epoch": 1.811125485122898,
2007
+ "grad_norm": 0.37963709235191345,
2008
+ "learning_rate": 2.058409570724842e-05,
2009
+ "loss": 0.3197,
2010
+ "step": 5600
2011
+ },
2012
+ {
2013
+ "epoch": 1.8175937904269082,
2014
+ "grad_norm": 0.4362301230430603,
2015
+ "learning_rate": 1.988036593947924e-05,
2016
+ "loss": 0.3405,
2017
+ "step": 5620
2018
+ },
2019
+ {
2020
+ "epoch": 1.8240620957309184,
2021
+ "grad_norm": 0.3662102520465851,
2022
+ "learning_rate": 1.9176636171710062e-05,
2023
+ "loss": 0.3262,
2024
+ "step": 5640
2025
+ },
2026
+ {
2027
+ "epoch": 1.8305304010349288,
2028
+ "grad_norm": 0.3674371838569641,
2029
+ "learning_rate": 1.847290640394089e-05,
2030
+ "loss": 0.3394,
2031
+ "step": 5660
2032
+ },
2033
+ {
2034
+ "epoch": 1.8369987063389392,
2035
+ "grad_norm": 0.31638914346694946,
2036
+ "learning_rate": 1.7769176636171712e-05,
2037
+ "loss": 0.3072,
2038
+ "step": 5680
2039
+ },
2040
+ {
2041
+ "epoch": 1.8434670116429497,
2042
+ "grad_norm": 0.39859539270401,
2043
+ "learning_rate": 1.7065446868402535e-05,
2044
+ "loss": 0.3398,
2045
+ "step": 5700
2046
+ },
2047
+ {
2048
+ "epoch": 1.84993531694696,
2049
+ "grad_norm": 0.37528204917907715,
2050
+ "learning_rate": 1.6361717100633358e-05,
2051
+ "loss": 0.3362,
2052
+ "step": 5720
2053
+ },
2054
+ {
2055
+ "epoch": 1.85640362225097,
2056
+ "grad_norm": 0.40589115023612976,
2057
+ "learning_rate": 1.565798733286418e-05,
2058
+ "loss": 0.3369,
2059
+ "step": 5740
2060
+ },
2061
+ {
2062
+ "epoch": 1.8628719275549805,
2063
+ "grad_norm": 0.3010273277759552,
2064
+ "learning_rate": 1.4954257565095003e-05,
2065
+ "loss": 0.3293,
2066
+ "step": 5760
2067
+ },
2068
+ {
2069
+ "epoch": 1.869340232858991,
2070
+ "grad_norm": 0.47433990240097046,
2071
+ "learning_rate": 1.4250527797325827e-05,
2072
+ "loss": 0.3137,
2073
+ "step": 5780
2074
+ },
2075
+ {
2076
+ "epoch": 1.8758085381630014,
2077
+ "grad_norm": 0.3495091199874878,
2078
+ "learning_rate": 1.354679802955665e-05,
2079
+ "loss": 0.3077,
2080
+ "step": 5800
2081
+ },
2082
+ {
2083
+ "epoch": 1.8822768434670116,
2084
+ "grad_norm": 0.33797210454940796,
2085
+ "learning_rate": 1.2843068261787475e-05,
2086
+ "loss": 0.3454,
2087
+ "step": 5820
2088
+ },
2089
+ {
2090
+ "epoch": 1.8887451487710218,
2091
+ "grad_norm": 0.34069564938545227,
2092
+ "learning_rate": 1.2139338494018297e-05,
2093
+ "loss": 0.3087,
2094
+ "step": 5840
2095
+ },
2096
+ {
2097
+ "epoch": 1.8952134540750323,
2098
+ "grad_norm": 0.4089032709598541,
2099
+ "learning_rate": 1.1435608726249122e-05,
2100
+ "loss": 0.3487,
2101
+ "step": 5860
2102
+ },
2103
+ {
2104
+ "epoch": 1.9016817593790427,
2105
+ "grad_norm": 0.3203008472919464,
2106
+ "learning_rate": 1.0731878958479945e-05,
2107
+ "loss": 0.3301,
2108
+ "step": 5880
2109
+ },
2110
+ {
2111
+ "epoch": 1.9081500646830531,
2112
+ "grad_norm": 0.3505829870700836,
2113
+ "learning_rate": 1.0028149190710768e-05,
2114
+ "loss": 0.3319,
2115
+ "step": 5900
2116
+ },
2117
+ {
2118
+ "epoch": 1.9146183699870634,
2119
+ "grad_norm": 0.29925537109375,
2120
+ "learning_rate": 9.324419422941591e-06,
2121
+ "loss": 0.3317,
2122
+ "step": 5920
2123
+ },
2124
+ {
2125
+ "epoch": 1.9210866752910736,
2126
+ "grad_norm": 0.4060503840446472,
2127
+ "learning_rate": 8.620689655172414e-06,
2128
+ "loss": 0.3393,
2129
+ "step": 5940
2130
+ },
2131
+ {
2132
+ "epoch": 1.927554980595084,
2133
+ "grad_norm": 0.4205261170864105,
2134
+ "learning_rate": 7.916959887403237e-06,
2135
+ "loss": 0.3375,
2136
+ "step": 5960
2137
+ },
2138
+ {
2139
+ "epoch": 1.9340232858990944,
2140
+ "grad_norm": 0.44757845997810364,
2141
+ "learning_rate": 7.213230119634061e-06,
2142
+ "loss": 0.3329,
2143
+ "step": 5980
2144
+ },
2145
+ {
2146
+ "epoch": 1.9404915912031049,
2147
+ "grad_norm": 0.5326984524726868,
2148
+ "learning_rate": 6.509500351864884e-06,
2149
+ "loss": 0.342,
2150
+ "step": 6000
2151
+ },
2152
+ {
2153
+ "epoch": 1.9404915912031049,
2154
+ "eval_loss": 0.5839738845825195,
2155
+ "eval_runtime": 967.7859,
2156
+ "eval_samples_per_second": 6.331,
2157
+ "eval_steps_per_second": 0.791,
2158
+ "step": 6000
2159
+ },
2160
+ {
2161
+ "epoch": 1.946959896507115,
2162
+ "grad_norm": 0.36838072538375854,
2163
+ "learning_rate": 5.8057705840957075e-06,
2164
+ "loss": 0.3123,
2165
+ "step": 6020
2166
+ },
2167
+ {
2168
+ "epoch": 1.9534282018111255,
2169
+ "grad_norm": 0.34141698479652405,
2170
+ "learning_rate": 5.102040816326531e-06,
2171
+ "loss": 0.3292,
2172
+ "step": 6040
2173
+ },
2174
+ {
2175
+ "epoch": 1.9598965071151357,
2176
+ "grad_norm": 0.3648461103439331,
2177
+ "learning_rate": 4.398311048557354e-06,
2178
+ "loss": 0.3078,
2179
+ "step": 6060
2180
+ },
2181
+ {
2182
+ "epoch": 1.9663648124191462,
2183
+ "grad_norm": 0.32175323367118835,
2184
+ "learning_rate": 3.6945812807881777e-06,
2185
+ "loss": 0.3033,
2186
+ "step": 6080
2187
+ },
2188
+ {
2189
+ "epoch": 1.9728331177231566,
2190
+ "grad_norm": 0.31717827916145325,
2191
+ "learning_rate": 2.990851513019001e-06,
2192
+ "loss": 0.3028,
2193
+ "step": 6100
2194
+ },
2195
+ {
2196
+ "epoch": 1.9793014230271668,
2197
+ "grad_norm": 0.5603770613670349,
2198
+ "learning_rate": 2.287121745249824e-06,
2199
+ "loss": 0.3414,
2200
+ "step": 6120
2201
+ },
2202
+ {
2203
+ "epoch": 1.9857697283311773,
2204
+ "grad_norm": 0.36055296659469604,
2205
+ "learning_rate": 1.5833919774806475e-06,
2206
+ "loss": 0.3202,
2207
+ "step": 6140
2208
+ },
2209
+ {
2210
+ "epoch": 1.9922380336351875,
2211
+ "grad_norm": 0.43246975541114807,
2212
+ "learning_rate": 8.796622097114708e-07,
2213
+ "loss": 0.3382,
2214
+ "step": 6160
2215
+ },
2216
+ {
2217
+ "epoch": 1.998706338939198,
2218
+ "grad_norm": 0.3947717249393463,
2219
+ "learning_rate": 1.7593244194229417e-07,
2220
+ "loss": 0.318,
2221
+ "step": 6180
2222
+ }
2223
+ ],
2224
+ "logging_steps": 20,
2225
+ "max_steps": 6184,
2226
+ "num_input_tokens_seen": 0,
2227
+ "num_train_epochs": 2,
2228
+ "save_steps": 1000,
2229
+ "stateful_callbacks": {
2230
+ "TrainerControl": {
2231
+ "args": {
2232
+ "should_epoch_stop": false,
2233
+ "should_evaluate": false,
2234
+ "should_log": false,
2235
+ "should_save": true,
2236
+ "should_training_stop": true
2237
+ },
2238
+ "attributes": {}
2239
+ }
2240
+ },
2241
+ "total_flos": 8.235501163189567e+18,
2242
+ "train_batch_size": 8,
2243
+ "trial_name": null,
2244
+ "trial_params": null
2245
+ }
ultival_mistral/checkpoint-6184/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8e5db938f70ff2ddfe787ade649d14bb55cc2be7a65bd63a20c3303bf68492
3
+ size 5368
ultival_mistral/runs/Jul03_16-12-32_a183f886ff78/events.out.tfevents.1751559215.a183f886ff78.1482.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b120d0ce6d01806a33ad7290440a5809ea4d1b07201b096bee144699018b9b8a
3
+ size 72589