jeromeramos commited on
Commit
fccb816
·
verified ·
1 Parent(s): 8910228

Model save

Browse files
README.md CHANGED
@@ -1,9 +1,7 @@
1
  ---
2
  base_model: Sim4Rec/inter-play-sim-assistant-sft
3
- datasets:
4
- - Sim4Rec/dpo_data
5
  library_name: transformers
6
- model_name: Sim4Rec/inter-play-sim-assistant-sft
7
  tags:
8
  - generated_from_trainer
9
  - trl
@@ -11,9 +9,9 @@ tags:
11
  licence: license
12
  ---
13
 
14
- # Model Card for Sim4Rec/inter-play-sim-assistant-sft
15
 
16
- This model is a fine-tuned version of [Sim4Rec/inter-play-sim-assistant-sft](https://huggingface.co/Sim4Rec/inter-play-sim-assistant-sft) on the [['Sim4Rec/dpo_data']](https://huggingface.co/datasets/['Sim4Rec/dpo_data']) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jerome-ramos-20/huggingface/runs/bw9lk46t)
33
 
34
 
35
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
 
1
  ---
2
  base_model: Sim4Rec/inter-play-sim-assistant-sft
 
 
3
  library_name: transformers
4
+ model_name: inter-play-sim-assistant-dpo
5
  tags:
6
  - generated_from_trainer
7
  - trl
 
9
  licence: license
10
  ---
11
 
12
+ # Model Card for inter-play-sim-assistant-dpo
13
 
14
+ This model is a fine-tuned version of [Sim4Rec/inter-play-sim-assistant-sft](https://huggingface.co/Sim4Rec/inter-play-sim-assistant-sft).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jerome-ramos-20/huggingface/runs/q3zfrxdb)
31
 
32
 
33
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 1.9938900203665988,
3
  "total_flos": 0.0,
4
- "train_loss": 0.5645458033827485,
5
- "train_runtime": 476.0531,
6
- "train_samples": 3921,
7
- "train_samples_per_second": 16.473,
8
- "train_steps_per_second": 0.256
9
  }
 
1
  {
2
+ "epoch": 2.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.22515048312672067,
5
+ "train_runtime": 2958.2669,
6
+ "train_samples": 45561,
7
+ "train_samples_per_second": 30.802,
8
+ "train_steps_per_second": 0.241
9
  }
config.json CHANGED
@@ -32,6 +32,6 @@
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
  "transformers_version": "4.48.2",
35
- "use_cache": true,
36
  "vocab_size": 128320
37
  }
 
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
  "transformers_version": "4.48.2",
35
+ "use_cache": false,
36
  "vocab_size": 128320
37
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c662bff069e16a55c0ed3957b7c1a7e2e7b6efe4c7dc1b38a09a42e95da5419d
3
  size 4977222960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228fda64e183c9fdc3a7708fb3208210cdadad23990e8923b3fe182ecf5c0cec
3
  size 4977222960
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e41f89c0b9e9c5a5e8dd10e5c19c527e4726dc9b373296f7a897cbbaa737621
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f7204be8a0fc4f037c035587cfd68a41793f7158bf6cf5794a16e4fe64cbc8a
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6e57a5558dd5fcdae6c05b45dd4d4a6294dd49f4b1dd9e5a5bba7cf612cffe6
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b841d9240b04b4dfe85f97b3b566448cfbabe276eac748a7ad65411df8577f
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c466502280c55a6dd02d9a75fc8f3c83841694b9133c35db128ed464184cfa86
3
  size 1168663096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c785dc0e0713c891a93b159479d9de5baf67a041e01652bf3e46d23a5251c8c
3
  size 1168663096
runs/Feb05_22-06-55_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf/events.out.tfevents.1738793347.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf.16929.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60bd33b78c7b1b2eb67898c2239db968a826c5b936caca18188d3c546ef10b8e
3
+ size 105020
runs/Feb05_23-51-38_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf/events.out.tfevents.1738799532.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf.60419.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95c3abbe38290f51d753e412d52ddc5244ed9e9d32021e7c0d13d40fbb8c6015
3
+ size 6438
runs/Feb05_23-52-52_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf/events.out.tfevents.1738799601.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5csnwcf.62482.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0069be6e747818e3901feb93b381cfe03ea449b10d5f618874b380e666cc459f
3
+ size 7790
runs/Feb06_00-02-58_w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5fcdrjv/events.out.tfevents.1738800332.w-jerom-inter-play-sim-94c6890b9ccf44ea86f033a3db8a5dbd-5fcdrjv.8855.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cc1728ddee65f74857ea0bc72b128753b118fe8fd026ac833ebb3d5af24cd87
3
+ size 56172
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 1.9938900203665988,
3
  "total_flos": 0.0,
4
- "train_loss": 0.5645458033827485,
5
- "train_runtime": 476.0531,
6
- "train_samples": 3921,
7
- "train_samples_per_second": 16.473,
8
- "train_steps_per_second": 0.256
9
  }
 
1
  {
2
+ "epoch": 2.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.22515048312672067,
5
+ "train_runtime": 2958.2669,
6
+ "train_samples": 45561,
7
+ "train_samples_per_second": 30.802,
8
+ "train_steps_per_second": 0.241
9
  }
trainer_state.json CHANGED
@@ -1,21 +1,21 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.9938900203665988,
5
  "eval_steps": 500,
6
- "global_step": 122,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.016293279022403257,
13
- "grad_norm": 1.4111297130584717,
14
- "learning_rate": 3.846153846153846e-08,
15
- "logits/chosen": -3.2578125,
16
- "logits/rejected": -3.19140625,
17
- "logps/chosen": -46.375,
18
- "logps/rejected": -45.75,
19
  "loss": 0.6914,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
@@ -24,197 +24,1082 @@
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.1629327902240326,
28
- "grad_norm": 0.7594385147094727,
29
- "learning_rate": 3.8461538461538463e-07,
30
- "logits/chosen": -3.2265625,
31
- "logits/rejected": -3.2200520038604736,
32
- "logps/chosen": -42.77604293823242,
33
- "logps/rejected": -41.88541793823242,
34
- "loss": 0.6924,
35
- "rewards/accuracies": 0.2222222238779068,
36
- "rewards/chosen": 0.0023810069542378187,
37
- "rewards/margins": -0.0003809928894042969,
38
- "rewards/rejected": 0.0027594566345214844,
39
  "step": 10
40
  },
41
  {
42
- "epoch": 0.3258655804480652,
43
- "grad_norm": 0.7135093808174133,
44
- "learning_rate": 4.949291683053768e-07,
45
- "logits/chosen": -3.2763671875,
46
- "logits/rejected": -3.2466797828674316,
47
- "logps/chosen": -40.720314025878906,
48
- "logps/rejected": -39.4296875,
49
- "loss": 0.6872,
50
- "rewards/accuracies": 0.421875,
51
- "rewards/chosen": 0.033158015459775925,
52
- "rewards/margins": 0.012033844366669655,
53
- "rewards/rejected": 0.02114267274737358,
54
  "step": 20
55
  },
56
  {
57
- "epoch": 0.48879837067209775,
58
- "grad_norm": 0.699894368648529,
59
- "learning_rate": 4.70586371748506e-07,
60
- "logits/chosen": -3.2841796875,
61
- "logits/rejected": -3.2529296875,
62
- "logps/chosen": -40.8515625,
63
- "logps/rejected": -40.571876525878906,
64
- "loss": 0.6765,
65
- "rewards/accuracies": 0.528124988079071,
66
- "rewards/chosen": 0.0396418571472168,
67
- "rewards/margins": 0.034914396703243256,
68
- "rewards/rejected": 0.004716300871223211,
69
  "step": 30
70
  },
71
  {
72
- "epoch": 0.6517311608961304,
73
- "grad_norm": 1.1233805418014526,
74
- "learning_rate": 4.280458575653296e-07,
75
- "logits/chosen": -3.195117235183716,
76
- "logits/rejected": -3.162890672683716,
77
- "logps/chosen": -43.59375,
78
- "logps/rejected": -44.532814025878906,
79
- "loss": 0.6415,
80
- "rewards/accuracies": 0.628125011920929,
81
- "rewards/chosen": -0.06923361122608185,
82
- "rewards/margins": 0.11445312201976776,
83
- "rewards/rejected": -0.1837112456560135,
84
  "step": 40
85
  },
86
  {
87
- "epoch": 0.814663951120163,
88
- "grad_norm": 1.2118208408355713,
89
- "learning_rate": 3.7081709127108767e-07,
90
- "logits/chosen": -3.0589842796325684,
91
  "logits/rejected": NaN,
92
- "logps/chosen": -50.13593673706055,
93
- "logps/rejected": -52.04999923706055,
94
- "loss": 0.6138,
95
- "rewards/accuracies": 0.612500011920929,
96
- "rewards/chosen": -0.3750244081020355,
97
- "rewards/margins": 0.18903808295726776,
98
- "rewards/rejected": -0.5643554925918579,
99
  "step": 50
100
  },
101
  {
102
- "epoch": 0.9775967413441955,
103
- "grad_norm": 1.2914516925811768,
104
- "learning_rate": 3.0362127536287636e-07,
105
- "logits/chosen": -3.135546922683716,
106
- "logits/rejected": -3.056835889816284,
107
- "logps/chosen": -50.14531326293945,
108
- "logps/rejected": -56.23749923706055,
109
- "loss": 0.566,
110
- "rewards/accuracies": 0.612500011920929,
111
- "rewards/chosen": -0.4345153868198395,
112
- "rewards/margins": 0.36333543062210083,
113
- "rewards/rejected": -0.7979736328125,
114
  "step": 60
115
  },
116
  {
117
- "epoch": 1.1466395112016294,
118
- "grad_norm": 1.4273715019226074,
119
- "learning_rate": 2.3200186419770823e-07,
120
- "logits/chosen": -3.161651134490967,
121
- "logits/rejected": -3.0788965225219727,
122
- "logps/chosen": -53.49691390991211,
123
- "logps/rejected": -62.85802459716797,
124
- "loss": 0.5371,
125
- "rewards/accuracies": 0.6095678806304932,
126
- "rewards/chosen": -0.5622889995574951,
127
- "rewards/margins": 0.5249747037887573,
128
- "rewards/rejected": -1.0871431827545166,
129
  "step": 70
130
  },
131
  {
132
- "epoch": 1.309572301425662,
133
- "grad_norm": 1.3244953155517578,
134
- "learning_rate": 1.6186724554503237e-07,
135
- "logits/chosen": -3.133984327316284,
136
- "logits/rejected": -3.0191407203674316,
137
- "logps/chosen": -55.234375,
138
- "logps/rejected": -68.234375,
139
- "loss": 0.4905,
140
- "rewards/accuracies": 0.660937488079071,
141
- "rewards/chosen": -0.6110439300537109,
142
- "rewards/margins": 0.7057861089706421,
143
- "rewards/rejected": -1.3170897960662842,
144
  "step": 80
145
  },
146
  {
147
- "epoch": 1.4725050916496945,
148
- "grad_norm": 1.8561619520187378,
149
- "learning_rate": 9.900331622138063e-08,
150
- "logits/chosen": -3.107617139816284,
151
- "logits/rejected": -2.9839844703674316,
152
- "logps/chosen": -55.421875,
153
- "logps/rejected": -68.80937194824219,
154
- "loss": 0.4936,
155
- "rewards/accuracies": 0.640625,
156
- "rewards/chosen": -0.6593307256698608,
157
- "rewards/margins": 0.737597644329071,
158
- "rewards/rejected": -1.3976562023162842,
159
  "step": 90
160
  },
161
  {
162
- "epoch": 1.635437881873727,
163
- "grad_norm": 0.9447304606437683,
164
- "learning_rate": 4.859616286322094e-08,
165
- "logits/chosen": -3.1148438453674316,
166
- "logits/rejected": -2.9876952171325684,
167
- "logps/chosen": -53.092185974121094,
168
- "logps/rejected": -68.9312515258789,
169
- "loss": 0.468,
170
- "rewards/accuracies": 0.6484375,
171
- "rewards/chosen": -0.5302764773368835,
172
- "rewards/margins": 0.887438952922821,
173
- "rewards/rejected": -1.417944312095642,
174
  "step": 100
175
  },
176
  {
177
- "epoch": 1.7983706720977597,
178
- "grad_norm": 1.593487024307251,
179
- "learning_rate": 1.4804225250339281e-08,
180
- "logits/chosen": -3.1207032203674316,
181
- "logits/rejected": NaN,
182
- "logps/chosen": -54.428123474121094,
183
- "logps/rejected": -68.76249694824219,
184
- "loss": 0.4758,
185
- "rewards/accuracies": 0.640625,
186
- "rewards/chosen": -0.6090973019599915,
187
- "rewards/margins": 0.817614734172821,
188
- "rewards/rejected": -1.4268066883087158,
189
  "step": 110
190
  },
191
  {
192
- "epoch": 1.9613034623217924,
193
- "grad_norm": 2.265634059906006,
194
- "learning_rate": 4.152374292708538e-10,
195
- "logits/chosen": -3.109179735183716,
196
- "logits/rejected": -2.985156297683716,
197
- "logps/chosen": -54.99687576293945,
198
- "logps/rejected": -71.015625,
199
- "loss": 0.4654,
200
- "rewards/accuracies": 0.667187511920929,
201
- "rewards/chosen": -0.6471847295761108,
202
- "rewards/margins": 0.863818347454071,
203
- "rewards/rejected": -1.510644555091858,
204
  "step": 120
205
  },
206
  {
207
- "epoch": 1.9938900203665988,
208
- "step": 122,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
209
  "total_flos": 0.0,
210
- "train_loss": 0.5645458033827485,
211
- "train_runtime": 476.0531,
212
- "train_samples_per_second": 16.473,
213
- "train_steps_per_second": 0.256
214
  }
215
  ],
216
  "logging_steps": 10,
217
- "max_steps": 122,
218
  "num_input_tokens_seen": 0,
219
  "num_train_epochs": 2,
220
  "save_steps": 500,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
  "eval_steps": 500,
6
+ "global_step": 712,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.0028089887640449437,
13
+ "grad_norm": 1.1993484497070312,
14
+ "learning_rate": 6.9444444444444435e-09,
15
+ "logits/chosen": -3.220703125,
16
+ "logits/rejected": -3.1796875,
17
+ "logps/chosen": -43.75,
18
+ "logps/rejected": -42.78125,
19
  "loss": 0.6914,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
 
24
  "step": 1
25
  },
26
  {
27
+ "epoch": 0.028089887640449437,
28
+ "grad_norm": 0.7986037135124207,
29
+ "learning_rate": 6.944444444444444e-08,
30
+ "logits/chosen": -3.2486979961395264,
31
+ "logits/rejected": -3.2052950859069824,
32
+ "logps/chosen": -42.34375,
33
+ "logps/rejected": -41.01215362548828,
34
+ "loss": 0.6921,
35
+ "rewards/accuracies": 0.2161458283662796,
36
+ "rewards/chosen": 0.00020620558643713593,
37
+ "rewards/margins": -0.00015253490710165352,
38
+ "rewards/rejected": 0.0003587139945011586,
39
  "step": 10
40
  },
41
  {
42
+ "epoch": 0.056179775280898875,
43
+ "grad_norm": 0.6845986247062683,
44
+ "learning_rate": 1.3888888888888888e-07,
45
+ "logits/chosen": -3.244140625,
46
+ "logits/rejected": -3.219531297683716,
47
+ "logps/chosen": -41.37812423706055,
48
+ "logps/rejected": -40.42656326293945,
49
+ "loss": 0.6907,
50
+ "rewards/accuracies": 0.3460937440395355,
51
+ "rewards/chosen": 0.006803750991821289,
52
+ "rewards/margins": 0.003222918603569269,
53
+ "rewards/rejected": 0.003580451011657715,
54
  "step": 20
55
  },
56
  {
57
+ "epoch": 0.08426966292134831,
58
+ "grad_norm": 0.5183084607124329,
59
+ "learning_rate": 2.0833333333333333e-07,
60
+ "logits/chosen": -3.263671875,
61
+ "logits/rejected": -3.2261719703674316,
62
+ "logps/chosen": -41.142189025878906,
63
+ "logps/rejected": -40.00312423706055,
64
+ "loss": 0.6862,
65
+ "rewards/accuracies": 0.44140625,
66
+ "rewards/chosen": 0.02759246900677681,
67
+ "rewards/margins": 0.013646435923874378,
68
+ "rewards/rejected": 0.013950538821518421,
69
  "step": 30
70
  },
71
  {
72
+ "epoch": 0.11235955056179775,
73
+ "grad_norm": 0.5848017334938049,
74
+ "learning_rate": 2.7777777777777776e-07,
75
+ "logits/chosen": -3.286328077316284,
76
+ "logits/rejected": -3.229296922683716,
77
+ "logps/chosen": -41.439064025878906,
78
+ "logps/rejected": -41.0,
79
+ "loss": 0.6745,
80
+ "rewards/accuracies": 0.5546875,
81
+ "rewards/chosen": 0.039928339421749115,
82
+ "rewards/margins": 0.038716744631528854,
83
+ "rewards/rejected": 0.001224255538545549,
84
  "step": 40
85
  },
86
  {
87
+ "epoch": 0.1404494382022472,
88
+ "grad_norm": 0.7922030091285706,
89
+ "learning_rate": 3.472222222222222e-07,
90
+ "logits/chosen": -3.2728514671325684,
91
  "logits/rejected": NaN,
92
+ "logps/chosen": -41.71406173706055,
93
+ "logps/rejected": -41.890625,
94
+ "loss": 0.6535,
95
+ "rewards/accuracies": 0.645312488079071,
96
+ "rewards/chosen": 0.03761863708496094,
97
+ "rewards/margins": 0.08368835598230362,
98
+ "rewards/rejected": -0.04610452800989151,
99
  "step": 50
100
  },
101
  {
102
+ "epoch": 0.16853932584269662,
103
+ "grad_norm": 1.3905400037765503,
104
+ "learning_rate": 4.1666666666666667e-07,
105
+ "logits/chosen": -3.204296827316284,
106
+ "logits/rejected": -3.154296875,
107
+ "logps/chosen": -44.493751525878906,
108
+ "logps/rejected": -48.446876525878906,
109
+ "loss": 0.5839,
110
+ "rewards/accuracies": 0.692187488079071,
111
+ "rewards/chosen": -0.11696071922779083,
112
+ "rewards/margins": 0.25516968965530396,
113
+ "rewards/rejected": -0.3720703125,
114
  "step": 60
115
  },
116
  {
117
+ "epoch": 0.19662921348314608,
118
+ "grad_norm": 0.8105548024177551,
119
+ "learning_rate": 4.861111111111111e-07,
120
+ "logits/chosen": -3.23828125,
121
+ "logits/rejected": -3.130859375,
122
+ "logps/chosen": -47.71875,
123
+ "logps/rejected": -59.140625,
124
+ "loss": 0.4754,
125
+ "rewards/accuracies": 0.7093750238418579,
126
+ "rewards/chosen": -0.2639709413051605,
127
+ "rewards/margins": 0.6321045160293579,
128
+ "rewards/rejected": -0.895751953125,
129
  "step": 70
130
  },
131
  {
132
+ "epoch": 0.2247191011235955,
133
+ "grad_norm": 0.6645036339759827,
134
+ "learning_rate": 4.998072590601808e-07,
135
+ "logits/chosen": -3.2529296875,
136
+ "logits/rejected": NaN,
137
+ "logps/chosen": -43.765625,
138
+ "logps/rejected": -72.40937805175781,
139
+ "loss": 0.3394,
140
+ "rewards/accuracies": 0.73828125,
141
+ "rewards/chosen": -0.08822021633386612,
142
+ "rewards/margins": 1.5029785633087158,
143
+ "rewards/rejected": -1.591284155845642,
144
  "step": 80
145
  },
146
  {
147
+ "epoch": 0.25280898876404495,
148
+ "grad_norm": 0.436211496591568,
149
+ "learning_rate": 4.990247583129217e-07,
150
+ "logits/chosen": -3.268749952316284,
151
+ "logits/rejected": -2.9375,
152
+ "logps/chosen": -46.046875,
153
+ "logps/rejected": -89.46875,
154
+ "loss": 0.2745,
155
+ "rewards/accuracies": 0.7476562261581421,
156
+ "rewards/chosen": -0.22807636857032776,
157
+ "rewards/margins": 2.234375,
158
+ "rewards/rejected": -2.463183641433716,
159
  "step": 90
160
  },
161
  {
162
+ "epoch": 0.2808988764044944,
163
+ "grad_norm": 0.43823757767677307,
164
+ "learning_rate": 4.976423351108942e-07,
165
+ "logits/chosen": -3.2183594703674316,
166
+ "logits/rejected": NaN,
167
+ "logps/chosen": -49.96875,
168
+ "logps/rejected": -105.32499694824219,
169
+ "loss": 0.2356,
170
+ "rewards/accuracies": 0.75390625,
171
+ "rewards/chosen": -0.3900146484375,
172
+ "rewards/margins": 2.8427734375,
173
+ "rewards/rejected": -3.232421875,
174
  "step": 100
175
  },
176
  {
177
+ "epoch": 0.3089887640449438,
178
+ "grad_norm": 0.6719679832458496,
179
+ "learning_rate": 4.95663319832678e-07,
180
+ "logits/chosen": -3.2197265625,
181
+ "logits/rejected": -2.762890577316284,
182
+ "logps/chosen": -46.25312423706055,
183
+ "logps/rejected": -114.41874694824219,
184
+ "loss": 0.236,
185
+ "rewards/accuracies": 0.7523437738418579,
186
+ "rewards/chosen": -0.22943687438964844,
187
+ "rewards/margins": 3.4632811546325684,
188
+ "rewards/rejected": -3.6929688453674316,
189
  "step": 110
190
  },
191
  {
192
+ "epoch": 0.33707865168539325,
193
+ "grad_norm": 0.3029685914516449,
194
+ "learning_rate": 4.930924800994191e-07,
195
+ "logits/chosen": -3.2417969703674316,
196
+ "logits/rejected": -2.752734422683716,
197
+ "logps/chosen": -45.203125,
198
+ "logps/rejected": -123.83125305175781,
199
+ "loss": 0.1955,
200
+ "rewards/accuracies": 0.77734375,
201
+ "rewards/chosen": -0.130183607339859,
202
+ "rewards/margins": 4.001757621765137,
203
+ "rewards/rejected": -4.1328125,
204
  "step": 120
205
  },
206
  {
207
+ "epoch": 0.3651685393258427,
208
+ "grad_norm": 0.5936453938484192,
209
+ "learning_rate": 4.899360092892144e-07,
210
+ "logits/chosen": -3.2191405296325684,
211
+ "logits/rejected": -2.7181639671325684,
212
+ "logps/chosen": -46.69843673706055,
213
+ "logps/rejected": -128.28750610351562,
214
+ "loss": 0.2039,
215
+ "rewards/accuracies": 0.7679687738418579,
216
+ "rewards/chosen": -0.2604345381259918,
217
+ "rewards/margins": 4.139843940734863,
218
+ "rewards/rejected": -4.399609565734863,
219
+ "step": 130
220
+ },
221
+ {
222
+ "epoch": 0.39325842696629215,
223
+ "grad_norm": 0.4349266588687897,
224
+ "learning_rate": 4.862015116167195e-07,
225
+ "logits/chosen": -3.243945360183716,
226
+ "logits/rejected": NaN,
227
+ "logps/chosen": -45.373435974121094,
228
+ "logps/rejected": -133.71249389648438,
229
+ "loss": 0.1869,
230
+ "rewards/accuracies": 0.7789062261581421,
231
+ "rewards/chosen": -0.10051727294921875,
232
+ "rewards/margins": 4.529296875,
233
+ "rewards/rejected": -4.631054878234863,
234
+ "step": 140
235
+ },
236
+ {
237
+ "epoch": 0.42134831460674155,
238
+ "grad_norm": 0.32329249382019043,
239
+ "learning_rate": 4.81897983813931e-07,
240
+ "logits/chosen": -3.2386717796325684,
241
+ "logits/rejected": -2.702929735183716,
242
+ "logps/chosen": -46.21562576293945,
243
+ "logps/rejected": -138.9375,
244
+ "loss": 0.1833,
245
+ "rewards/accuracies": 0.780468761920929,
246
+ "rewards/chosen": -0.20498999953269958,
247
+ "rewards/margins": 4.690625190734863,
248
+ "rewards/rejected": -4.895898342132568,
249
+ "step": 150
250
+ },
251
+ {
252
+ "epoch": 0.449438202247191,
253
+ "grad_norm": 0.24011513590812683,
254
+ "learning_rate": 4.770357934562704e-07,
255
+ "logits/chosen": -3.281445264816284,
256
+ "logits/rejected": NaN,
257
+ "logps/chosen": -41.82500076293945,
258
+ "logps/rejected": -132.2624969482422,
259
+ "loss": 0.195,
260
+ "rewards/accuracies": 0.765625,
261
+ "rewards/chosen": 0.00346546177752316,
262
+ "rewards/margins": 4.6083984375,
263
+ "rewards/rejected": -4.605273246765137,
264
+ "step": 160
265
+ },
266
+ {
267
+ "epoch": 0.47752808988764045,
268
+ "grad_norm": 0.5602438449859619,
269
+ "learning_rate": 4.716266539861866e-07,
270
+ "logits/chosen": -3.237499952316284,
271
+ "logits/rejected": -2.6552734375,
272
+ "logps/chosen": -47.9921875,
273
+ "logps/rejected": -144.77499389648438,
274
+ "loss": 0.1834,
275
+ "rewards/accuracies": 0.7757812738418579,
276
+ "rewards/chosen": -0.31496095657348633,
277
+ "rewards/margins": 4.906640529632568,
278
+ "rewards/rejected": -5.219336032867432,
279
+ "step": 170
280
+ },
281
+ {
282
+ "epoch": 0.5056179775280899,
283
+ "grad_norm": 0.5013967156410217,
284
+ "learning_rate": 4.6568359649444796e-07,
285
+ "logits/chosen": -3.225781202316284,
286
+ "logits/rejected": -2.6431641578674316,
287
+ "logps/chosen": -47.4140625,
288
+ "logps/rejected": -147.6062469482422,
289
+ "loss": 0.1777,
290
+ "rewards/accuracies": 0.7835937738418579,
291
+ "rewards/chosen": -0.26249465346336365,
292
+ "rewards/margins": 5.070898532867432,
293
+ "rewards/rejected": -5.333398342132568,
294
+ "step": 180
295
+ },
296
+ {
297
+ "epoch": 0.5337078651685393,
298
+ "grad_norm": 0.3109186887741089,
299
+ "learning_rate": 4.592209383271023e-07,
300
+ "logits/chosen": -3.2769532203674316,
301
+ "logits/rejected": -2.708203077316284,
302
+ "logps/chosen": -40.32500076293945,
303
+ "logps/rejected": -146.55624389648438,
304
+ "loss": 0.1691,
305
+ "rewards/accuracies": 0.7945312261581421,
306
+ "rewards/chosen": 0.10864410549402237,
307
+ "rewards/margins": 5.372851371765137,
308
+ "rewards/rejected": -5.262499809265137,
309
+ "step": 190
310
+ },
311
+ {
312
+ "epoch": 0.5617977528089888,
313
+ "grad_norm": 0.329673707485199,
314
+ "learning_rate": 4.5225424859373684e-07,
315
+ "logits/chosen": -3.244921922683716,
316
+ "logits/rejected": -2.625195264816284,
317
+ "logps/chosen": -46.1875,
318
+ "logps/rejected": -152.55624389648438,
319
+ "loss": 0.1722,
320
+ "rewards/accuracies": 0.784375011920929,
321
+ "rewards/chosen": -0.20301513373851776,
322
+ "rewards/margins": 5.3759765625,
323
+ "rewards/rejected": -5.578711032867432,
324
+ "step": 200
325
+ },
326
+ {
327
+ "epoch": 0.5898876404494382,
328
+ "grad_norm": 0.6182069182395935,
329
+ "learning_rate": 4.448003106601291e-07,
330
+ "logits/chosen": -3.2255859375,
331
+ "logits/rejected": NaN,
332
+ "logps/chosen": -45.904685974121094,
333
+ "logps/rejected": -153.14999389648438,
334
+ "loss": 0.1818,
335
+ "rewards/accuracies": 0.7789062261581421,
336
+ "rewards/chosen": -0.2197813093662262,
337
+ "rewards/margins": 5.441601753234863,
338
+ "rewards/rejected": -5.661913871765137,
339
+ "step": 210
340
+ },
341
+ {
342
+ "epoch": 0.6179775280898876,
343
+ "grad_norm": 0.10240374505519867,
344
+ "learning_rate": 4.3687708171564917e-07,
345
+ "logits/chosen": -3.2201170921325684,
346
+ "logits/rejected": NaN,
347
+ "logps/chosen": -46.890625,
348
+ "logps/rejected": -154.24374389648438,
349
+ "loss": 0.18,
350
+ "rewards/accuracies": 0.76953125,
351
+ "rewards/chosen": -0.2859039306640625,
352
+ "rewards/margins": 5.441992282867432,
353
+ "rewards/rejected": -5.730078220367432,
354
+ "step": 220
355
+ },
356
+ {
357
+ "epoch": 0.6460674157303371,
358
+ "grad_norm": 0.5484092235565186,
359
+ "learning_rate": 4.2850364951281705e-07,
360
+ "logits/chosen": -3.2359375953674316,
361
+ "logits/rejected": -2.595507860183716,
362
+ "logps/chosen": -44.52656173706055,
363
+ "logps/rejected": -156.5187530517578,
364
+ "loss": 0.1763,
365
+ "rewards/accuracies": 0.785937488079071,
366
+ "rewards/chosen": -0.10360870510339737,
367
+ "rewards/margins": 5.668359279632568,
368
+ "rewards/rejected": -5.7724609375,
369
+ "step": 230
370
+ },
371
+ {
372
+ "epoch": 0.6741573033707865,
373
+ "grad_norm": 0.33201783895492554,
374
+ "learning_rate": 4.1970018638323547e-07,
375
+ "logits/chosen": -3.2476563453674316,
376
+ "logits/rejected": -2.6099610328674316,
377
+ "logps/chosen": -41.240623474121094,
378
+ "logps/rejected": -152.3249969482422,
379
+ "loss": 0.1836,
380
+ "rewards/accuracies": 0.778124988079071,
381
+ "rewards/chosen": 0.05380706861615181,
382
+ "rewards/margins": 5.612500190734863,
383
+ "rewards/rejected": -5.559960842132568,
384
+ "step": 240
385
+ },
386
+ {
387
+ "epoch": 0.702247191011236,
388
+ "grad_norm": 0.3225279450416565,
389
+ "learning_rate": 4.1048790064067573e-07,
390
+ "logits/chosen": -3.2152342796325684,
391
+ "logits/rejected": NaN,
392
+ "logps/chosen": -42.44062423706055,
393
+ "logps/rejected": -152.5437469482422,
394
+ "loss": 0.1874,
395
+ "rewards/accuracies": 0.7679687738418579,
396
+ "rewards/chosen": -0.019408416002988815,
397
+ "rewards/margins": 5.590234279632568,
398
+ "rewards/rejected": -5.608788967132568,
399
+ "step": 250
400
+ },
401
+ {
402
+ "epoch": 0.7303370786516854,
403
+ "grad_norm": 0.5013627409934998,
404
+ "learning_rate": 4.0088898548839285e-07,
405
+ "logits/chosen": -3.1996092796325684,
406
+ "logits/rejected": -2.558398485183716,
407
+ "logps/chosen": -46.775001525878906,
408
+ "logps/rejected": -157.64999389648438,
409
+ "loss": 0.1809,
410
+ "rewards/accuracies": 0.78125,
411
+ "rewards/chosen": -0.24601340293884277,
412
+ "rewards/margins": 5.5927734375,
413
+ "rewards/rejected": -5.837695121765137,
414
+ "step": 260
415
+ },
416
+ {
417
+ "epoch": 0.7584269662921348,
418
+ "grad_norm": 0.2542474567890167,
419
+ "learning_rate": 3.9092656555375414e-07,
420
+ "logits/chosen": -3.184765577316284,
421
+ "logits/rejected": -2.549999952316284,
422
+ "logps/chosen": -48.16718673706055,
423
+ "logps/rejected": -155.3937530517578,
424
+ "loss": 0.1927,
425
+ "rewards/accuracies": 0.7593749761581421,
426
+ "rewards/chosen": -0.3481277525424957,
427
+ "rewards/margins": 5.426171779632568,
428
+ "rewards/rejected": -5.775000095367432,
429
+ "step": 270
430
+ },
431
+ {
432
+ "epoch": 0.7865168539325843,
433
+ "grad_norm": 1.486909031867981,
434
+ "learning_rate": 3.806246411789872e-07,
435
+ "logits/chosen": -3.206835985183716,
436
+ "logits/rejected": -2.568554639816284,
437
+ "logps/chosen": -46.279685974121094,
438
+ "logps/rejected": -161.4499969482422,
439
+ "loss": 0.1695,
440
+ "rewards/accuracies": 0.796875,
441
+ "rewards/chosen": -0.17198029160499573,
442
+ "rewards/margins": 5.812304496765137,
443
+ "rewards/rejected": -5.983984470367432,
444
+ "step": 280
445
+ },
446
+ {
447
+ "epoch": 0.8146067415730337,
448
+ "grad_norm": 0.24136996269226074,
449
+ "learning_rate": 3.700080306022528e-07,
450
+ "logits/chosen": -3.238085985183716,
451
+ "logits/rejected": -2.606250047683716,
452
+ "logps/chosen": -42.40312576293945,
453
+ "logps/rejected": -155.80624389648438,
454
+ "loss": 0.1873,
455
+ "rewards/accuracies": 0.76953125,
456
+ "rewards/chosen": -0.016921233385801315,
457
+ "rewards/margins": 5.745703220367432,
458
+ "rewards/rejected": -5.7607421875,
459
+ "step": 290
460
+ },
461
+ {
462
+ "epoch": 0.8426966292134831,
463
+ "grad_norm": 0.3395313620567322,
464
+ "learning_rate": 3.5910231016833546e-07,
465
+ "logits/chosen": -3.2119140625,
466
+ "logits/rejected": NaN,
467
+ "logps/chosen": -45.810935974121094,
468
+ "logps/rejected": -161.1687469482422,
469
+ "loss": 0.174,
470
+ "rewards/accuracies": 0.78125,
471
+ "rewards/chosen": -0.18533477187156677,
472
+ "rewards/margins": 5.834374904632568,
473
+ "rewards/rejected": -6.019140720367432,
474
+ "step": 300
475
+ },
476
+ {
477
+ "epoch": 0.8707865168539326,
478
+ "grad_norm": 0.14939536154270172,
479
+ "learning_rate": 3.4793375271298895e-07,
480
+ "logits/chosen": -3.1988282203674316,
481
+ "logits/rejected": -2.597851514816284,
482
+ "logps/chosen": -44.76874923706055,
483
+ "logps/rejected": -163.0437469482422,
484
+ "loss": 0.1656,
485
+ "rewards/accuracies": 0.7867187261581421,
486
+ "rewards/chosen": -0.11381302028894424,
487
+ "rewards/margins": 5.981835842132568,
488
+ "rewards/rejected": -6.096875190734863,
489
+ "step": 310
490
+ },
491
+ {
492
+ "epoch": 0.898876404494382,
493
+ "grad_norm": 0.6819726228713989,
494
+ "learning_rate": 3.3652926426937325e-07,
495
+ "logits/chosen": -3.246875047683716,
496
+ "logits/rejected": -2.609179735183716,
497
+ "logps/chosen": -41.875,
498
+ "logps/rejected": -162.28125,
499
+ "loss": 0.1824,
500
+ "rewards/accuracies": 0.7828124761581421,
501
+ "rewards/chosen": 0.06156463548541069,
502
+ "rewards/margins": 6.103515625,
503
+ "rewards/rejected": -6.043359279632568,
504
+ "step": 320
505
+ },
506
+ {
507
+ "epoch": 0.9269662921348315,
508
+ "grad_norm": 0.10508494079113007,
509
+ "learning_rate": 3.249163192490642e-07,
510
+ "logits/chosen": -3.2544922828674316,
511
+ "logits/rejected": -2.624804735183716,
512
+ "logps/chosen": -41.25,
513
+ "logps/rejected": -159.97500610351562,
514
+ "loss": 0.1816,
515
+ "rewards/accuracies": 0.7734375,
516
+ "rewards/chosen": 0.021668624132871628,
517
+ "rewards/margins": 5.989648342132568,
518
+ "rewards/rejected": -5.967968940734863,
519
+ "step": 330
520
+ },
521
+ {
522
+ "epoch": 0.9550561797752809,
523
+ "grad_norm": 0.2005680650472641,
524
+ "learning_rate": 3.1312289425378944e-07,
525
+ "logits/chosen": -3.2759766578674316,
526
+ "logits/rejected": -2.6624999046325684,
527
+ "logps/chosen": -39.45781326293945,
528
+ "logps/rejected": -157.4375,
529
+ "loss": 0.1757,
530
+ "rewards/accuracies": 0.78125,
531
+ "rewards/chosen": 0.14164963364601135,
532
+ "rewards/margins": 5.972851753234863,
533
+ "rewards/rejected": -5.831250190734863,
534
+ "step": 340
535
+ },
536
+ {
537
+ "epoch": 0.9831460674157303,
538
+ "grad_norm": 0.2797287404537201,
539
+ "learning_rate": 3.011774006773449e-07,
540
+ "logits/chosen": -3.282031297683716,
541
+ "logits/rejected": NaN,
542
+ "logps/chosen": -37.98906326293945,
543
+ "logps/rejected": -157.4375,
544
+ "loss": 0.1713,
545
+ "rewards/accuracies": 0.7757812738418579,
546
+ "rewards/chosen": 0.19436034560203552,
547
+ "rewards/margins": 6.061718940734863,
548
+ "rewards/rejected": -5.866406440734863,
549
+ "step": 350
550
+ },
551
+ {
552
+ "epoch": 1.0112359550561798,
553
+ "grad_norm": 0.1386144906282425,
554
+ "learning_rate": 2.8910861626005773e-07,
555
+ "logits/chosen": -3.2330079078674316,
556
+ "logits/rejected": -2.6128907203674316,
557
+ "logps/chosen": -42.69062423706055,
558
+ "logps/rejected": -162.75625610351562,
559
+ "loss": 0.1689,
560
+ "rewards/accuracies": 0.7942708730697632,
561
+ "rewards/chosen": -0.019468307495117188,
562
+ "rewards/margins": 6.076367378234863,
563
+ "rewards/rejected": -6.09765625,
564
+ "step": 360
565
+ },
566
+ {
567
+ "epoch": 1.0393258426966292,
568
+ "grad_norm": 0.31265443563461304,
569
+ "learning_rate": 2.7694561576068983e-07,
570
+ "logits/chosen": -3.234375,
571
+ "logits/rejected": -2.5785155296325684,
572
+ "logps/chosen": -41.896873474121094,
573
+ "logps/rejected": -160.8125,
574
+ "loss": 0.1713,
575
+ "rewards/accuracies": 0.7867187261581421,
576
+ "rewards/chosen": -0.01103897113353014,
577
+ "rewards/margins": 6.00390625,
578
+ "rewards/rejected": -6.014843940734863,
579
+ "step": 370
580
+ },
581
+ {
582
+ "epoch": 1.0674157303370786,
583
+ "grad_norm": 0.3972986042499542,
584
+ "learning_rate": 2.647177009127972e-07,
585
+ "logits/chosen": -3.2544922828674316,
586
+ "logits/rejected": -2.5933594703674316,
587
+ "logps/chosen": -39.5703125,
588
+ "logps/rejected": -160.85000610351562,
589
+ "loss": 0.1793,
590
+ "rewards/accuracies": 0.768750011920929,
591
+ "rewards/chosen": 0.09815521538257599,
592
+ "rewards/margins": 6.125781059265137,
593
+ "rewards/rejected": -6.029492378234863,
594
+ "step": 380
595
+ },
596
+ {
597
+ "epoch": 1.095505617977528,
598
+ "grad_norm": 0.12721529603004456,
599
+ "learning_rate": 2.524543298342874e-07,
600
+ "logits/chosen": -3.262890577316284,
601
+ "logits/rejected": -2.592578172683716,
602
+ "logps/chosen": -40.4765625,
603
+ "logps/rejected": -161.91250610351562,
604
+ "loss": 0.1743,
605
+ "rewards/accuracies": 0.77734375,
606
+ "rewards/chosen": 0.05222644656896591,
607
+ "rewards/margins": 6.142578125,
608
+ "rewards/rejected": -6.091406345367432,
609
+ "step": 390
610
+ },
611
+ {
612
+ "epoch": 1.1235955056179776,
613
+ "grad_norm": 0.09501045197248459,
614
+ "learning_rate": 2.401850460602329e-07,
615
+ "logits/chosen": -3.2220702171325684,
616
+ "logits/rejected": -2.559765577316284,
617
+ "logps/chosen": -44.423439025878906,
618
+ "logps/rejected": -167.08749389648438,
619
+ "loss": 0.1718,
620
+ "rewards/accuracies": 0.780468761920929,
621
+ "rewards/chosen": -0.09567908942699432,
622
+ "rewards/margins": 6.20703125,
623
+ "rewards/rejected": -6.302734375,
624
+ "step": 400
625
+ },
626
+ {
627
+ "epoch": 1.151685393258427,
628
+ "grad_norm": 0.29580923914909363,
629
+ "learning_rate": 2.2793940736990766e-07,
630
+ "logits/chosen": -3.2007813453674316,
631
+ "logits/rejected": NaN,
632
+ "logps/chosen": -51.12968826293945,
633
+ "logps/rejected": -170.96875,
634
+ "loss": 0.172,
635
+ "rewards/accuracies": 0.780468761920929,
636
+ "rewards/chosen": -0.42823487520217896,
637
+ "rewards/margins": 6.079297065734863,
638
+ "rewards/rejected": -6.508593559265137,
639
+ "step": 410
640
+ },
641
+ {
642
+ "epoch": 1.1797752808988764,
643
+ "grad_norm": 0.2765229046344757,
644
+ "learning_rate": 2.1574691457950803e-07,
645
+ "logits/chosen": -3.1839842796325684,
646
+ "logits/rejected": NaN,
647
+ "logps/chosen": -50.9609375,
648
+ "logps/rejected": -170.78125,
649
+ "loss": 0.1704,
650
+ "rewards/accuracies": 0.7734375,
651
+ "rewards/chosen": -0.46327972412109375,
652
+ "rewards/margins": 6.071484565734863,
653
+ "rewards/rejected": -6.534375190734863,
654
+ "step": 420
655
+ },
656
+ {
657
+ "epoch": 1.2078651685393258,
658
+ "grad_norm": 0.15300235152244568,
659
+ "learning_rate": 2.036369404721023e-07,
660
+ "logits/chosen": -3.2113280296325684,
661
+ "logits/rejected": -2.531054735183716,
662
+ "logps/chosen": -46.8828125,
663
+ "logps/rejected": -173.1374969482422,
664
+ "loss": 0.1599,
665
+ "rewards/accuracies": 0.79296875,
666
+ "rewards/chosen": -0.21221771836280823,
667
+ "rewards/margins": 6.374609470367432,
668
+ "rewards/rejected": -6.587109565734863,
669
+ "step": 430
670
+ },
671
+ {
672
+ "epoch": 1.2359550561797752,
673
+ "grad_norm": 0.6692606806755066,
674
+ "learning_rate": 1.9163865903602372e-07,
675
+ "logits/chosen": -3.253124952316284,
676
+ "logits/rejected": -2.5308594703674316,
677
+ "logps/chosen": -42.287498474121094,
678
+ "logps/rejected": -169.625,
679
+ "loss": 0.1762,
680
+ "rewards/accuracies": 0.776562511920929,
681
+ "rewards/chosen": -0.029553985223174095,
682
+ "rewards/margins": 6.40625,
683
+ "rewards/rejected": -6.434765815734863,
684
+ "step": 440
685
+ },
686
+ {
687
+ "epoch": 1.2640449438202248,
688
+ "grad_norm": 0.2745380103588104,
689
+ "learning_rate": 1.7978097518217702e-07,
690
+ "logits/chosen": -3.237499952316284,
691
+ "logits/rejected": -2.541015625,
692
+ "logps/chosen": -42.74687576293945,
693
+ "logps/rejected": -168.78750610351562,
694
+ "loss": 0.1715,
695
+ "rewards/accuracies": 0.774218738079071,
696
+ "rewards/chosen": -0.021244239062070847,
697
+ "rewards/margins": 6.392187595367432,
698
+ "rewards/rejected": -6.409375190734863,
699
+ "step": 450
700
+ },
701
+ {
702
+ "epoch": 1.2921348314606742,
703
+ "grad_norm": 0.3858237564563751,
704
+ "learning_rate": 1.6809245510957666e-07,
705
+ "logits/chosen": -3.235546827316284,
706
+ "logits/rejected": NaN,
707
+ "logps/chosen": -44.904685974121094,
708
+ "logps/rejected": -172.96249389648438,
709
+ "loss": 0.1663,
710
+ "rewards/accuracies": 0.7875000238418579,
711
+ "rewards/chosen": -0.10267486423254013,
712
+ "rewards/margins": 6.484375,
713
+ "rewards/rejected": -6.587109565734863,
714
+ "step": 460
715
+ },
716
+ {
717
+ "epoch": 1.3202247191011236,
718
+ "grad_norm": 0.25496408343315125,
719
+ "learning_rate": 1.5660125748687093e-07,
720
+ "logits/chosen": -3.224414110183716,
721
+ "logits/rejected": NaN,
722
+ "logps/chosen": -46.64374923706055,
723
+ "logps/rejected": -171.1374969482422,
724
+ "loss": 0.1703,
725
+ "rewards/accuracies": 0.780468761920929,
726
+ "rewards/chosen": -0.2242431640625,
727
+ "rewards/margins": 6.318749904632568,
728
+ "rewards/rejected": -6.542187690734863,
729
+ "step": 470
730
+ },
731
+ {
732
+ "epoch": 1.348314606741573,
733
+ "grad_norm": 0.3984578251838684,
734
+ "learning_rate": 1.4533506561564305e-07,
735
+ "logits/chosen": -3.224609375,
736
+ "logits/rejected": -2.529101610183716,
737
+ "logps/chosen": -47.29375076293945,
738
+ "logps/rejected": -173.3625030517578,
739
+ "loss": 0.1669,
740
+ "rewards/accuracies": 0.780468761920929,
741
+ "rewards/chosen": -0.2605232298374176,
742
+ "rewards/margins": 6.368359565734863,
743
+ "rewards/rejected": -6.627734184265137,
744
+ "step": 480
745
+ },
746
+ {
747
+ "epoch": 1.3764044943820224,
748
+ "grad_norm": 0.39370596408843994,
749
+ "learning_rate": 1.343210207389125e-07,
750
+ "logits/chosen": -3.203320264816284,
751
+ "logits/rejected": -2.5302734375,
752
+ "logps/chosen": -48.0859375,
753
+ "logps/rejected": -172.93124389648438,
754
+ "loss": 0.1747,
755
+ "rewards/accuracies": 0.7757812738418579,
756
+ "rewards/chosen": -0.27991026639938354,
757
+ "rewards/margins": 6.329297065734863,
758
+ "rewards/rejected": -6.610156059265137,
759
+ "step": 490
760
+ },
761
+ {
762
+ "epoch": 1.404494382022472,
763
+ "grad_norm": 0.2013687640428543,
764
+ "learning_rate": 1.2358565665550387e-07,
765
+ "logits/chosen": -3.244335889816284,
766
+ "logits/rejected": NaN,
767
+ "logps/chosen": -41.96875,
768
+ "logps/rejected": -176.03750610351562,
769
+ "loss": 0.1557,
770
+ "rewards/accuracies": 0.793749988079071,
771
+ "rewards/chosen": 0.01244263630360365,
772
+ "rewards/margins": 6.784765720367432,
773
+ "rewards/rejected": -6.769921779632568,
774
+ "step": 500
775
+ },
776
+ {
777
+ "epoch": 1.4325842696629214,
778
+ "grad_norm": 0.24023930728435516,
779
+ "learning_rate": 1.1315483579780094e-07,
780
+ "logits/chosen": -3.2484374046325684,
781
+ "logits/rejected": -2.5347657203674316,
782
+ "logps/chosen": -42.046875,
783
+ "logps/rejected": -175.8312530517578,
784
+ "loss": 0.1643,
785
+ "rewards/accuracies": 0.788281261920929,
786
+ "rewards/chosen": 0.040112875401973724,
787
+ "rewards/margins": 6.779296875,
788
+ "rewards/rejected": -6.739453315734863,
789
+ "step": 510
790
+ },
791
+ {
792
+ "epoch": 1.4606741573033708,
793
+ "grad_norm": 0.5066425800323486,
794
+ "learning_rate": 1.0305368692688174e-07,
795
+ "logits/chosen": -3.240429639816284,
796
+ "logits/rejected": -2.522656202316284,
797
+ "logps/chosen": -43.00468826293945,
798
+ "logps/rejected": -173.8625030517578,
799
+ "loss": 0.1791,
800
+ "rewards/accuracies": 0.778124988079071,
801
+ "rewards/chosen": -0.03853149339556694,
802
+ "rewards/margins": 6.608788967132568,
803
+ "rewards/rejected": -6.646874904632568,
804
+ "step": 520
805
+ },
806
+ {
807
+ "epoch": 1.4887640449438202,
808
+ "grad_norm": 0.10324753075838089,
809
+ "learning_rate": 9.330654459513266e-08,
810
+ "logits/chosen": -3.233203172683716,
811
+ "logits/rejected": -2.5269532203674316,
812
+ "logps/chosen": -41.953125,
813
+ "logps/rejected": -172.06875610351562,
814
+ "loss": 0.1683,
815
+ "rewards/accuracies": 0.7749999761581421,
816
+ "rewards/chosen": -0.02672729454934597,
817
+ "rewards/margins": 6.571875095367432,
818
+ "rewards/rejected": -6.598437309265137,
819
+ "step": 530
820
+ },
821
+ {
822
+ "epoch": 1.5168539325842696,
823
+ "grad_norm": 0.06884948909282684,
824
+ "learning_rate": 8.393689052217964e-08,
825
+ "logits/chosen": -3.2310547828674316,
826
+ "logits/rejected": -2.5240235328674316,
827
+ "logps/chosen": -41.89531326293945,
828
+ "logps/rejected": -170.4499969482422,
829
+ "loss": 0.1774,
830
+ "rewards/accuracies": 0.770312488079071,
831
+ "rewards/chosen": -0.03569946438074112,
832
+ "rewards/margins": 6.483984470367432,
833
+ "rewards/rejected": -6.521093845367432,
834
+ "step": 540
835
+ },
836
+ {
837
+ "epoch": 1.5449438202247192,
838
+ "grad_norm": 0.2964838445186615,
839
+ "learning_rate": 7.49672970253691e-08,
840
+ "logits/chosen": -3.2562499046325684,
841
+ "logits/rejected": -2.5296874046325684,
842
+ "logps/chosen": -40.365623474121094,
843
+ "logps/rejected": -173.3625030517578,
844
+ "loss": 0.1644,
845
+ "rewards/accuracies": 0.78515625,
846
+ "rewards/chosen": 0.06430435180664062,
847
+ "rewards/margins": 6.696484565734863,
848
+ "rewards/rejected": -6.630078315734863,
849
+ "step": 550
850
+ },
851
+ {
852
+ "epoch": 1.5730337078651684,
853
+ "grad_norm": 0.16073837876319885,
854
+ "learning_rate": 6.641937264107867e-08,
855
+ "logits/chosen": -3.2535157203674316,
856
+ "logits/rejected": -2.5425782203674316,
857
+ "logps/chosen": -39.43281173706055,
858
+ "logps/rejected": -167.91250610351562,
859
+ "loss": 0.1788,
860
+ "rewards/accuracies": 0.7671874761581421,
861
+ "rewards/chosen": 0.08403320610523224,
862
+ "rewards/margins": 6.482421875,
863
+ "rewards/rejected": -6.396874904632568,
864
+ "step": 560
865
+ },
866
+ {
867
+ "epoch": 1.601123595505618,
868
+ "grad_norm": 0.27828249335289,
869
+ "learning_rate": 5.831371006785962e-08,
870
+ "logits/chosen": -3.2699217796325684,
871
+ "logits/rejected": -2.5228514671325684,
872
+ "logps/chosen": -39.87187576293945,
873
+ "logps/rejected": -171.1437530517578,
874
+ "loss": 0.175,
875
+ "rewards/accuracies": 0.772656261920929,
876
+ "rewards/chosen": 0.07085514068603516,
877
+ "rewards/margins": 6.617578029632568,
878
+ "rewards/rejected": -6.548047065734863,
879
+ "step": 570
880
+ },
881
+ {
882
+ "epoch": 1.6292134831460674,
883
+ "grad_norm": 0.386394202709198,
884
+ "learning_rate": 5.066983655682325e-08,
885
+ "logits/chosen": -3.2455077171325684,
886
+ "logits/rejected": -2.546093702316284,
887
+ "logps/chosen": -40.72968673706055,
888
+ "logps/rejected": -168.77499389648438,
889
+ "loss": 0.1783,
890
+ "rewards/accuracies": 0.7679687738418579,
891
+ "rewards/chosen": 0.032080840319395065,
892
+ "rewards/margins": 6.479296684265137,
893
+ "rewards/rejected": -6.448828220367432,
894
+ "step": 580
895
+ },
896
+ {
897
+ "epoch": 1.6573033707865168,
898
+ "grad_norm": 0.3948213756084442,
899
+ "learning_rate": 4.3506166868781755e-08,
900
+ "logits/chosen": -3.2671875953674316,
901
+ "logits/rejected": -2.521484375,
902
+ "logps/chosen": -41.20624923706055,
903
+ "logps/rejected": -172.6875,
904
+ "loss": 0.1759,
905
+ "rewards/accuracies": 0.76953125,
906
+ "rewards/chosen": 0.012204742059111595,
907
+ "rewards/margins": 6.639843940734863,
908
+ "rewards/rejected": -6.629296779632568,
909
+ "step": 590
910
+ },
911
+ {
912
+ "epoch": 1.6853932584269664,
913
+ "grad_norm": 0.4554503560066223,
914
+ "learning_rate": 3.683995891147695e-08,
915
+ "logits/chosen": -3.246289014816284,
916
+ "logits/rejected": NaN,
917
+ "logps/chosen": -42.18281173706055,
918
+ "logps/rejected": -173.86874389648438,
919
+ "loss": 0.1625,
920
+ "rewards/accuracies": 0.784375011920929,
921
+ "rewards/chosen": 0.0017807006370276213,
922
+ "rewards/margins": 6.689453125,
923
+ "rewards/rejected": -6.688672065734863,
924
+ "step": 600
925
+ },
926
+ {
927
+ "epoch": 1.7134831460674156,
928
+ "grad_norm": 0.6775197386741638,
929
+ "learning_rate": 3.0687272163768986e-08,
930
+ "logits/chosen": -3.259960889816284,
931
+ "logits/rejected": NaN,
932
+ "logps/chosen": -41.376564025878906,
933
+ "logps/rejected": -174.4499969482422,
934
+ "loss": 0.1619,
935
+ "rewards/accuracies": 0.788281261920929,
936
+ "rewards/chosen": 0.07793807983398438,
937
+ "rewards/margins": 6.741796970367432,
938
+ "rewards/rejected": -6.6640625,
939
+ "step": 610
940
+ },
941
+ {
942
+ "epoch": 1.7415730337078652,
943
+ "grad_norm": 0.11740182340145111,
944
+ "learning_rate": 2.5062928986944676e-08,
945
+ "logits/chosen": -3.2728514671325684,
946
+ "logits/rejected": NaN,
947
+ "logps/chosen": -40.89374923706055,
948
+ "logps/rejected": -175.0625,
949
+ "loss": 0.1611,
950
+ "rewards/accuracies": 0.7835937738418579,
951
+ "rewards/chosen": 0.07378844916820526,
952
+ "rewards/margins": 6.797656059265137,
953
+ "rewards/rejected": -6.723828315734863,
954
+ "step": 620
955
+ },
956
+ {
957
+ "epoch": 1.7696629213483146,
958
+ "grad_norm": 0.23929810523986816,
959
+ "learning_rate": 1.9980478916351296e-08,
960
+ "logits/chosen": -3.2544922828674316,
961
+ "logits/rejected": -2.5478515625,
962
+ "logps/chosen": -42.296875,
963
+ "logps/rejected": -174.1374969482422,
964
+ "loss": 0.1689,
965
+ "rewards/accuracies": 0.785937488079071,
966
+ "rewards/chosen": -0.01357345562428236,
967
+ "rewards/margins": 6.648046970367432,
968
+ "rewards/rejected": -6.661328315734863,
969
+ "step": 630
970
+ },
971
+ {
972
+ "epoch": 1.797752808988764,
973
+ "grad_norm": 0.06252361834049225,
974
+ "learning_rate": 1.5452166019378987e-08,
975
+ "logits/chosen": -3.2582030296325684,
976
+ "logits/rejected": -2.5250000953674316,
977
+ "logps/chosen": -39.837501525878906,
978
+ "logps/rejected": -173.9875030517578,
979
+ "loss": 0.1693,
980
+ "rewards/accuracies": 0.776562511920929,
981
+ "rewards/chosen": 0.06287650763988495,
982
+ "rewards/margins": 6.756249904632568,
983
+ "rewards/rejected": -6.696875095367432,
984
+ "step": 640
985
+ },
986
+ {
987
+ "epoch": 1.8258426966292136,
988
+ "grad_norm": 0.6466670632362366,
989
+ "learning_rate": 1.1488899398429896e-08,
990
+ "logits/chosen": -3.2587890625,
991
+ "logits/rejected": -2.5443358421325684,
992
+ "logps/chosen": -41.12812423706055,
993
+ "logps/rejected": -175.3125,
994
+ "loss": 0.1643,
995
+ "rewards/accuracies": 0.793749988079071,
996
+ "rewards/chosen": 0.057323455810546875,
997
+ "rewards/margins": 6.766015529632568,
998
+ "rewards/rejected": -6.710156440734863,
999
+ "step": 650
1000
+ },
1001
+ {
1002
+ "epoch": 1.8539325842696628,
1003
+ "grad_norm": 0.364284485578537,
1004
+ "learning_rate": 8.100226909935059e-09,
1005
+ "logits/chosen": -3.2474608421325684,
1006
+ "logits/rejected": -2.5396485328674316,
1007
+ "logps/chosen": -43.040626525878906,
1008
+ "logps/rejected": -175.02499389648438,
1009
+ "loss": 0.1656,
1010
+ "rewards/accuracies": 0.7906249761581421,
1011
+ "rewards/chosen": -0.026231002062559128,
1012
+ "rewards/margins": 6.670312404632568,
1013
+ "rewards/rejected": -6.696093559265137,
1014
+ "step": 660
1015
+ },
1016
+ {
1017
+ "epoch": 1.8820224719101124,
1018
+ "grad_norm": 0.1942785233259201,
1019
+ "learning_rate": 5.2943121627319346e-09,
1020
+ "logits/chosen": -3.252734422683716,
1021
+ "logits/rejected": -2.544140577316284,
1022
+ "logps/chosen": -41.951560974121094,
1023
+ "logps/rejected": -170.24374389648438,
1024
+ "loss": 0.1726,
1025
+ "rewards/accuracies": 0.770312488079071,
1026
+ "rewards/chosen": -0.00887908972799778,
1027
+ "rewards/margins": 6.493359565734863,
1028
+ "rewards/rejected": -6.500390529632568,
1029
+ "step": 670
1030
+ },
1031
+ {
1032
+ "epoch": 1.9101123595505618,
1033
+ "grad_norm": 0.38263440132141113,
1034
+ "learning_rate": 3.077914851215585e-09,
1035
+ "logits/chosen": -3.257031202316284,
1036
+ "logits/rejected": -2.5390625,
1037
+ "logps/chosen": -41.865623474121094,
1038
+ "logps/rejected": -176.53750610351562,
1039
+ "loss": 0.1582,
1040
+ "rewards/accuracies": 0.796093761920929,
1041
+ "rewards/chosen": 0.024802017956972122,
1042
+ "rewards/margins": 6.796093940734863,
1043
+ "rewards/rejected": -6.771484375,
1044
+ "step": 680
1045
+ },
1046
+ {
1047
+ "epoch": 1.9382022471910112,
1048
+ "grad_norm": 0.4378679096698761,
1049
+ "learning_rate": 1.4563744706429514e-09,
1050
+ "logits/chosen": -3.2578125,
1051
+ "logits/rejected": -2.5419921875,
1052
+ "logps/chosen": -40.009376525878906,
1053
+ "logps/rejected": -171.91250610351562,
1054
+ "loss": 0.1703,
1055
+ "rewards/accuracies": 0.7749999761581421,
1056
+ "rewards/chosen": 0.07464599609375,
1057
+ "rewards/margins": 6.642968654632568,
1058
+ "rewards/rejected": -6.567968845367432,
1059
+ "step": 690
1060
+ },
1061
+ {
1062
+ "epoch": 1.9662921348314608,
1063
+ "grad_norm": 0.19188769161701202,
1064
+ "learning_rate": 4.3359745382104405e-10,
1065
+ "logits/chosen": -3.271679639816284,
1066
+ "logits/rejected": -2.5396485328674316,
1067
+ "logps/chosen": -41.52031326293945,
1068
+ "logps/rejected": -175.375,
1069
+ "loss": 0.1601,
1070
+ "rewards/accuracies": 0.7906249761581421,
1071
+ "rewards/chosen": 0.01084976177662611,
1072
+ "rewards/margins": 6.757421970367432,
1073
+ "rewards/rejected": -6.750390529632568,
1074
+ "step": 700
1075
+ },
1076
+ {
1077
+ "epoch": 1.99438202247191,
1078
+ "grad_norm": 0.2792583107948303,
1079
+ "learning_rate": 1.2047760167999133e-11,
1080
+ "logits/chosen": -3.2544922828674316,
1081
+ "logits/rejected": -2.5517578125,
1082
+ "logps/chosen": -42.29999923706055,
1083
+ "logps/rejected": -175.3625030517578,
1084
+ "loss": 0.1594,
1085
+ "rewards/accuracies": 0.7867187261581421,
1086
+ "rewards/chosen": -0.014897918328642845,
1087
+ "rewards/margins": 6.701171875,
1088
+ "rewards/rejected": -6.717187404632568,
1089
+ "step": 710
1090
+ },
1091
+ {
1092
+ "epoch": 2.0,
1093
+ "step": 712,
1094
  "total_flos": 0.0,
1095
+ "train_loss": 0.22515048312672067,
1096
+ "train_runtime": 2958.2669,
1097
+ "train_samples_per_second": 30.802,
1098
+ "train_steps_per_second": 0.241
1099
  }
1100
  ],
1101
  "logging_steps": 10,
1102
+ "max_steps": 712,
1103
  "num_input_tokens_seen": 0,
1104
  "num_train_epochs": 2,
1105
  "save_steps": 500,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6be99c816b5db8c5a3acc17c9a53f9de39b2335b62b1dba0929d841f39fa676d
3
  size 7800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e47b8bd96f5ed212913b9d45c5a069e52c690acee03a5d1528a81d6eb90ec43
3
  size 7800